diff --git a/v1.25/alvistack-vagrant-kubernetes/PRODUCT.yaml b/v1.25/alvistack-vagrant-kubernetes/PRODUCT.yaml index 8f9ebb5544..c1b409e97f 100644 --- a/v1.25/alvistack-vagrant-kubernetes/PRODUCT.yaml +++ b/v1.25/alvistack-vagrant-kubernetes/PRODUCT.yaml @@ -1,6 +1,6 @@ vendor: AlviStack name: AlviStack - Vagrant Box Packaging for Kubernetes -version: 20221213.1.1 +version: 20230225.1.1 website_url: https://github.com/alvistack/vagrant-kubernetes documentation_url: https://github.com/alvistack/ansible-collection-kubernetes/tree/master/docs product_logo_url: https://raw.githubusercontent.com/alvistack/ansible-collection-kubernetes/master/docs/images/alvistack/stacked/color/alvistack-stacked-color.svg diff --git a/v1.25/alvistack-vagrant-kubernetes/README.md b/v1.25/alvistack-vagrant-kubernetes/README.md index fe5f6ade9b..de5684d1bc 100644 --- a/v1.25/alvistack-vagrant-kubernetes/README.md +++ b/v1.25/alvistack-vagrant-kubernetes/README.md @@ -4,33 +4,33 @@ For running k8s conformance test we need 2 vagrant instances as master and 1 vagrant instance as node with following minimal system requirement, e.g. - - host - - libvirt - - nested virtualization enabled - - Ubuntu 22.04 - - 8 CPUs - - 32GB RAM - - `kube01` - - kubernetes master, etcd - - cri-o, flannel - - Ubuntu 22.04 - - IP: 192.168.121.101/24 - - 2 CPUs - - 8GB RAM - - `kube02` - - kubernetes master, etcd - - cri-o, flannel - - Ubuntu 22.04 - - IP: 192.168.121.102/24 - - 2 CPUs - - 8GB RAM - - `kube03` - - kubernetes node, etcd - - cri-o, flannel - - Ubuntu 22.04 - - IP: 192.168.121.103/24 - - 2 CPUs - - 8GB RAM +- host + - libvirt + - nested virtualization enabled + - Ubuntu 22.04 + - 8 CPUs + - 32GB RAM +- `kube01` + - kubernetes master, etcd + - cri-o, cilium + - Ubuntu 22.04 + - IP: 192.168.121.101/24 + - 2 CPUs + - 8GB RAM +- `kube02` + - kubernetes master, etcd + - cri-o, cilium + - Ubuntu 22.04 + - IP: 192.168.121.102/24 + - 2 CPUs + - 8GB RAM +- `kube03` + - kubernetes node, etcd + - cri-o, cilium + - Ubuntu 22.04 + - IP: 192.168.121.103/24 + - 2 CPUs + - 8GB RAM ## Bootstrap Host @@ -99,29 +99,33 @@ All instances could be SSH and switch as root with `sudo su -`, e.g. Check result: root@kube01:~# kubectl get node - NAME STATUS ROLES AGE VERSION - kube01 Ready control-plane,master 9m38s v1.25.5 - kube02 Ready control-plane,master 8m32s v1.25.5 - kube03 Ready 8m14s v1.25.5 - + NAME STATUS ROLES AGE VERSION + kube01 Ready control-plane 3h41m v1.25.6 + kube02 Ready control-plane 3h40m v1.25.6 + kube03 Ready 3h36m v1.25.6 + root@kube01:~# kubectl get pod --all-namespaces - NAMESPACE NAME READY STATUS RESTARTS AGE - kube-flannel kube-flannel-ds-5z9q7 1/1 Running 0 70s - kube-flannel kube-flannel-ds-7bf94 1/1 Running 0 70s - kube-flannel kube-flannel-ds-99tj8 1/1 Running 0 70s - kube-system coredns-bd6b6df9f-cqzkx 1/1 Running 0 70s - kube-system coredns-bd6b6df9f-tcv7m 1/1 Running 0 70s - kube-system kube-addon-manager-kube01 1/1 Running 2 80s - kube-system kube-addon-manager-kube02 1/1 Running 2 80s - kube-system kube-apiserver-kube01 1/1 Running 3 80s - kube-system kube-apiserver-kube02 1/1 Running 3 80s - kube-system kube-controller-manager-kube01 1/1 Running 3 80s - kube-system kube-controller-manager-kube02 1/1 Running 4 80s - kube-system kube-proxy-brwdv 1/1 Running 0 70s - kube-system kube-proxy-ppn7b 1/1 Running 0 70s - kube-system kube-proxy-qn7vg 1/1 Running 0 70s - kube-system kube-scheduler-kube01 1/1 Running 3 80s - kube-system kube-scheduler-kube02 1/1 Running 4 80s + NAMESPACE NAME READY STATUS RESTARTS AGE + kube-system cilium-2cncz 1/1 Running 0 3h36m + kube-system cilium-9tkfg 1/1 Running 0 3h39m + kube-system cilium-node-init-8b7n2 1/1 Running 0 3h39m + kube-system cilium-node-init-979mb 1/1 Running 0 3h36m + kube-system cilium-node-init-s84z5 1/1 Running 0 3h39m + kube-system cilium-operator-59d7d77cdc-lppbd 1/1 Running 0 3h39m + kube-system cilium-r4lc4 1/1 Running 0 3h39m + kube-system coredns-565d847f94-5nh2n 1/1 Running 0 3h38m + kube-system coredns-565d847f94-qqvf6 1/1 Running 0 3h38m + kube-system kube-addon-manager-kube01 1/1 Running 0 3h40m + kube-system kube-addon-manager-kube02 1/1 Running 0 3h40m + kube-system kube-apiserver-kube01 1/1 Running 0 3h41m + kube-system kube-apiserver-kube02 1/1 Running 0 3h40m + kube-system kube-controller-manager-kube01 1/1 Running 0 3h41m + kube-system kube-controller-manager-kube02 1/1 Running 0 3h40m + kube-system kube-proxy-22s9q 1/1 Running 0 3h36m + kube-system kube-proxy-cm4qg 1/1 Running 0 3h40m + kube-system kube-proxy-d66m7 1/1 Running 0 3h40m + kube-system kube-scheduler-kube01 1/1 Running 0 3h41m + kube-system kube-scheduler-kube02 1/1 Running 0 3h40m ## Run Sonobuoy @@ -129,7 +133,7 @@ Run sonobuoy for conformance test as official procedure (): root@kube01:~# sonobuoy run --mode=certified-conformance - + root@kube01:~# sonobuoy status - + root@kube01:~# sonobuoy retrieve diff --git a/v1.25/alvistack-vagrant-kubernetes/e2e.log b/v1.25/alvistack-vagrant-kubernetes/e2e.log index c960bd674f..53c6b72fd3 100644 --- a/v1.25/alvistack-vagrant-kubernetes/e2e.log +++ b/v1.25/alvistack-vagrant-kubernetes/e2e.log @@ -1,8 +1,8 @@ -I1214 15:36:18.137230 14 e2e.go:116] Starting e2e run "211b00e9-8cf9-4510-82bc-c1b7016a99c3" on Ginkgo node 1 -Dec 14 15:36:18.163: INFO: Enabling in-tree volume drivers +I0225 09:46:42.344732 13 e2e.go:116] Starting e2e run "c4f82c74-9faf-42fb-b115-044538789898" on Ginkgo node 1 +Feb 25 09:46:42.377: INFO: Enabling in-tree volume drivers Running Suite: Kubernetes e2e suite - /usr/local/bin ==================================================== -Random Seed: 1671032177 - will randomize all specs +Random Seed: 1677318401 - will randomize all specs Will run 362 of 7066 specs ------------------------------ @@ -11,14060 +11,13158 @@ test/e2e/e2e.go:76 [SynchronizedBeforeSuite] TOP-LEVEL test/e2e/e2e.go:76 {"msg":"Test Suite starting","completed":0,"skipped":0,"failed":0} -Dec 14 15:36:18.326: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 15:36:18.331: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable -Dec 14 15:36:18.362: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready -Dec 14 15:36:18.401: INFO: 13 / 13 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) -Dec 14 15:36:18.401: INFO: expected 2 pod replicas in namespace 'kube-system', 2 are Running and Ready. -Dec 14 15:36:18.401: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start -Dec 14 15:36:18.409: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kube-proxy' (0 seconds elapsed) -Dec 14 15:36:18.409: INFO: e2e test version: v1.25.5 -Dec 14 15:36:18.412: INFO: kube-apiserver version: v1.25.5 +Feb 25 09:46:42.548: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 09:46:42.560: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable +Feb 25 09:46:42.595: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready +Feb 25 09:46:42.654: INFO: 20 / 20 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) +Feb 25 09:46:42.654: INFO: expected 3 pod replicas in namespace 'kube-system', 3 are Running and Ready. +Feb 25 09:46:42.654: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start +Feb 25 09:46:42.674: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'cilium' (0 seconds elapsed) +Feb 25 09:46:42.674: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'cilium-node-init' (0 seconds elapsed) +Feb 25 09:46:42.674: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kube-proxy' (0 seconds elapsed) +Feb 25 09:46:42.674: INFO: e2e test version: v1.25.6 +Feb 25 09:46:42.676: INFO: kube-apiserver version: v1.25.6 [SynchronizedBeforeSuite] TOP-LEVEL test/e2e/e2e.go:76 -Dec 14 15:36:18.412: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 15:36:18.425: INFO: Cluster IP family: ipv4 +Feb 25 09:46:42.677: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 09:46:42.688: INFO: Cluster IP family: ipv4 ------------------------------ -[SynchronizedBeforeSuite] PASSED [0.100 seconds] +[SynchronizedBeforeSuite] PASSED [0.141 seconds] [SynchronizedBeforeSuite] test/e2e/e2e.go:76 Begin Captured GinkgoWriter Output >> [SynchronizedBeforeSuite] TOP-LEVEL test/e2e/e2e.go:76 - Dec 14 15:36:18.326: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 15:36:18.331: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable - Dec 14 15:36:18.362: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready - Dec 14 15:36:18.401: INFO: 13 / 13 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) - Dec 14 15:36:18.401: INFO: expected 2 pod replicas in namespace 'kube-system', 2 are Running and Ready. - Dec 14 15:36:18.401: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start - Dec 14 15:36:18.409: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kube-proxy' (0 seconds elapsed) - Dec 14 15:36:18.409: INFO: e2e test version: v1.25.5 - Dec 14 15:36:18.412: INFO: kube-apiserver version: v1.25.5 + Feb 25 09:46:42.548: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 09:46:42.560: INFO: Waiting up to 30m0s for all (but 0) nodes to be schedulable + Feb 25 09:46:42.595: INFO: Waiting up to 10m0s for all pods (need at least 0) in namespace 'kube-system' to be running and ready + Feb 25 09:46:42.654: INFO: 20 / 20 pods in namespace 'kube-system' are running and ready (0 seconds elapsed) + Feb 25 09:46:42.654: INFO: expected 3 pod replicas in namespace 'kube-system', 3 are Running and Ready. + Feb 25 09:46:42.654: INFO: Waiting up to 5m0s for all daemonsets in namespace 'kube-system' to start + Feb 25 09:46:42.674: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'cilium' (0 seconds elapsed) + Feb 25 09:46:42.674: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'cilium-node-init' (0 seconds elapsed) + Feb 25 09:46:42.674: INFO: 3 / 3 pods ready in namespace 'kube-system' in daemonset 'kube-proxy' (0 seconds elapsed) + Feb 25 09:46:42.674: INFO: e2e test version: v1.25.6 + Feb 25 09:46:42.676: INFO: kube-apiserver version: v1.25.6 [SynchronizedBeforeSuite] TOP-LEVEL test/e2e/e2e.go:76 - Dec 14 15:36:18.412: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 15:36:18.425: INFO: Cluster IP family: ipv4 + Feb 25 09:46:42.677: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 09:46:42.688: INFO: Cluster IP family: ipv4 << End Captured GinkgoWriter Output ------------------------------ SSSS ------------------------------ -[sig-storage] ConfigMap - should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:98 -[BeforeEach] [sig-storage] ConfigMap +[sig-storage] Downward API volume + should provide container's memory request [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:234 +[BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:36:18.453 -Dec 14 15:36:18.454: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename configmap 12/14/22 15:36:18.455 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:36:18.496 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:36:18.501 -[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:98 -STEP: Creating configMap with name configmap-test-volume-map-135a9925-9f16-4c78-a1ce-4f6b8effe168 12/14/22 15:36:18.505 -STEP: Creating a pod to test consume configMaps 12/14/22 15:36:18.514 -Dec 14 15:36:18.530: INFO: Waiting up to 5m0s for pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171" in namespace "configmap-542" to be "Succeeded or Failed" -Dec 14 15:36:18.534: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 3.887102ms -Dec 14 15:36:20.542: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011253551s -Dec 14 15:36:22.549: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 4.018460549s -Dec 14 15:36:24.546: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 6.015899851s -Dec 14 15:36:26.541: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 8.010166454s -Dec 14 15:36:28.540: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 10.009967682s -Dec 14 15:36:30.547: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 12.016275247s -Dec 14 15:36:32.540: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 14.010073923s -Dec 14 15:36:34.542: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 16.011224634s -Dec 14 15:36:36.540: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 18.009737133s -Dec 14 15:36:38.540: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 20.009794191s -Dec 14 15:36:40.542: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 22.011162527s -Dec 14 15:36:42.546: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 24.015455262s -Dec 14 15:36:44.544: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.013935714s -STEP: Saw pod success 12/14/22 15:36:44.545 -Dec 14 15:36:44.547: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171" satisfied condition "Succeeded or Failed" -Dec 14 15:36:44.553: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171 container agnhost-container: -STEP: delete the pod 12/14/22 15:36:44.585 -Dec 14 15:36:44.610: INFO: Waiting for pod pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171 to disappear -Dec 14 15:36:44.615: INFO: Pod pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171 no longer exists -[AfterEach] [sig-storage] ConfigMap +STEP: Creating a kubernetes client 02/25/23 09:46:42.723 +Feb 25 09:46:42.723: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 09:46:42.724 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:46:42.764 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:46:42.772 +[BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 +[It] should provide container's memory request [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:234 +STEP: Creating a pod to test downward API volume plugin 02/25/23 09:46:42.777 +Feb 25 09:46:42.797: INFO: Waiting up to 5m0s for pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c" in namespace "downward-api-1523" to be "Succeeded or Failed" +Feb 25 09:46:42.812: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 14.524687ms +Feb 25 09:46:44.818: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020544938s +Feb 25 09:46:46.819: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.022308172s +Feb 25 09:46:48.819: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 6.022500761s +Feb 25 09:46:50.822: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 8.025274564s +Feb 25 09:46:52.832: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 10.034599169s +Feb 25 09:46:54.820: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 12.022584079s +Feb 25 09:46:56.826: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 14.028685389s +Feb 25 09:46:58.819: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 16.021964445s +Feb 25 09:47:00.818: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 18.020996164s +Feb 25 09:47:02.822: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 20.025341061s +STEP: Saw pod success 02/25/23 09:47:02.822 +Feb 25 09:47:02.824: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c" satisfied condition "Succeeded or Failed" +Feb 25 09:47:02.832: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c container client-container: +STEP: delete the pod 02/25/23 09:47:02.865 +Feb 25 09:47:02.891: INFO: Waiting for pod downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c to disappear +Feb 25 09:47:02.897: INFO: Pod downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c no longer exists +[AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 -Dec 14 15:36:44.616: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-542" for this suite. 12/14/22 15:36:44.622 -{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","completed":1,"skipped":4,"failed":0} +Feb 25 09:47:02.897: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-1523" for this suite. 02/25/23 09:47:02.905 +{"msg":"PASSED [sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance]","completed":1,"skipped":4,"failed":0} ------------------------------ -• [SLOW TEST] [26.177 seconds] -[sig-storage] ConfigMap +• [SLOW TEST] [20.195 seconds] +[sig-storage] Downward API volume test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:98 + should provide container's memory request [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:234 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] ConfigMap + [BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:36:18.453 - Dec 14 15:36:18.454: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename configmap 12/14/22 15:36:18.455 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:36:18.496 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:36:18.501 - [It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:98 - STEP: Creating configMap with name configmap-test-volume-map-135a9925-9f16-4c78-a1ce-4f6b8effe168 12/14/22 15:36:18.505 - STEP: Creating a pod to test consume configMaps 12/14/22 15:36:18.514 - Dec 14 15:36:18.530: INFO: Waiting up to 5m0s for pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171" in namespace "configmap-542" to be "Succeeded or Failed" - Dec 14 15:36:18.534: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 3.887102ms - Dec 14 15:36:20.542: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011253551s - Dec 14 15:36:22.549: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 4.018460549s - Dec 14 15:36:24.546: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 6.015899851s - Dec 14 15:36:26.541: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 8.010166454s - Dec 14 15:36:28.540: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 10.009967682s - Dec 14 15:36:30.547: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 12.016275247s - Dec 14 15:36:32.540: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 14.010073923s - Dec 14 15:36:34.542: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 16.011224634s - Dec 14 15:36:36.540: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 18.009737133s - Dec 14 15:36:38.540: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 20.009794191s - Dec 14 15:36:40.542: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 22.011162527s - Dec 14 15:36:42.546: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Pending", Reason="", readiness=false. Elapsed: 24.015455262s - Dec 14 15:36:44.544: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.013935714s - STEP: Saw pod success 12/14/22 15:36:44.545 - Dec 14 15:36:44.547: INFO: Pod "pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171" satisfied condition "Succeeded or Failed" - Dec 14 15:36:44.553: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171 container agnhost-container: - STEP: delete the pod 12/14/22 15:36:44.585 - Dec 14 15:36:44.610: INFO: Waiting for pod pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171 to disappear - Dec 14 15:36:44.615: INFO: Pod pod-configmaps-663591ec-b7a4-49c5-a7ae-abba5e8a2171 no longer exists - [AfterEach] [sig-storage] ConfigMap + STEP: Creating a kubernetes client 02/25/23 09:46:42.723 + Feb 25 09:46:42.723: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 09:46:42.724 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:46:42.764 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:46:42.772 + [BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 + [It] should provide container's memory request [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:234 + STEP: Creating a pod to test downward API volume plugin 02/25/23 09:46:42.777 + Feb 25 09:46:42.797: INFO: Waiting up to 5m0s for pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c" in namespace "downward-api-1523" to be "Succeeded or Failed" + Feb 25 09:46:42.812: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 14.524687ms + Feb 25 09:46:44.818: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020544938s + Feb 25 09:46:46.819: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.022308172s + Feb 25 09:46:48.819: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 6.022500761s + Feb 25 09:46:50.822: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 8.025274564s + Feb 25 09:46:52.832: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 10.034599169s + Feb 25 09:46:54.820: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 12.022584079s + Feb 25 09:46:56.826: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 14.028685389s + Feb 25 09:46:58.819: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 16.021964445s + Feb 25 09:47:00.818: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Pending", Reason="", readiness=false. Elapsed: 18.020996164s + Feb 25 09:47:02.822: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 20.025341061s + STEP: Saw pod success 02/25/23 09:47:02.822 + Feb 25 09:47:02.824: INFO: Pod "downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c" satisfied condition "Succeeded or Failed" + Feb 25 09:47:02.832: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c container client-container: + STEP: delete the pod 02/25/23 09:47:02.865 + Feb 25 09:47:02.891: INFO: Waiting for pod downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c to disappear + Feb 25 09:47:02.897: INFO: Pod downwardapi-volume-aa3fd381-0ddc-4a7c-8900-11cb8fc5fb2c no longer exists + [AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 - Dec 14 15:36:44.616: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "configmap-542" for this suite. 12/14/22 15:36:44.622 + Feb 25 09:47:02.897: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-1523" for this suite. 02/25/23 09:47:02.905 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook - should execute poststart exec hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:97 -[BeforeEach] [sig-node] Container Lifecycle Hook +[sig-storage] Projected configMap + should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:98 +[BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:36:44.632 -Dec 14 15:36:44.632: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-lifecycle-hook 12/14/22 15:36:44.636 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:36:44.661 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:36:44.665 -[BeforeEach] when create a pod with lifecycle hook - test/e2e/common/node/lifecycle_hook.go:55 -STEP: create the container to handle the HTTPGet hook request. 12/14/22 15:36:44.676 -Dec 14 15:36:44.687: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-4607" to be "running and ready" -Dec 14 15:36:44.692: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 4.595169ms -Dec 14 15:36:44.692: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) -Dec 14 15:36:46.704: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 2.016801768s -Dec 14 15:36:46.704: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) -Dec 14 15:36:46.704: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" -[It] should execute poststart exec hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:97 -STEP: create the pod with lifecycle hook 12/14/22 15:36:46.711 -Dec 14 15:36:46.720: INFO: Waiting up to 5m0s for pod "pod-with-poststart-exec-hook" in namespace "container-lifecycle-hook-4607" to be "running and ready" -Dec 14 15:36:46.729: INFO: Pod "pod-with-poststart-exec-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 8.514546ms -Dec 14 15:36:46.729: INFO: The phase of Pod pod-with-poststart-exec-hook is Pending, waiting for it to be Running (with Ready = true) -Dec 14 15:36:48.737: INFO: Pod "pod-with-poststart-exec-hook": Phase="Running", Reason="", readiness=true. Elapsed: 2.016717908s -Dec 14 15:36:48.737: INFO: The phase of Pod pod-with-poststart-exec-hook is Running (Ready = true) -Dec 14 15:36:48.737: INFO: Pod "pod-with-poststart-exec-hook" satisfied condition "running and ready" -STEP: check poststart hook 12/14/22 15:36:48.741 -STEP: delete the pod with lifecycle hook 12/14/22 15:36:48.751 -Dec 14 15:36:48.759: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -Dec 14 15:36:48.769: INFO: Pod pod-with-poststart-exec-hook still exists -Dec 14 15:36:50.770: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -Dec 14 15:36:50.783: INFO: Pod pod-with-poststart-exec-hook still exists -Dec 14 15:36:52.770: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear -Dec 14 15:36:52.777: INFO: Pod pod-with-poststart-exec-hook no longer exists -[AfterEach] [sig-node] Container Lifecycle Hook +STEP: Creating a kubernetes client 02/25/23 09:47:02.933 +Feb 25 09:47:02.934: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 09:47:02.936 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:02.963 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:02.968 +[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:98 +STEP: Creating configMap with name projected-configmap-test-volume-map-80010730-b6b1-44d8-9efc-35c9ef6ee5ff 02/25/23 09:47:02.974 +STEP: Creating a pod to test consume configMaps 02/25/23 09:47:02.982 +Feb 25 09:47:02.997: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64" in namespace "projected-1842" to be "Succeeded or Failed" +Feb 25 09:47:03.006: INFO: Pod "pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64": Phase="Pending", Reason="", readiness=false. Elapsed: 8.95658ms +Feb 25 09:47:05.034: INFO: Pod "pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64": Phase="Pending", Reason="", readiness=false. Elapsed: 2.037508093s +Feb 25 09:47:07.014: INFO: Pod "pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.0174184s +STEP: Saw pod success 02/25/23 09:47:07.014 +Feb 25 09:47:07.015: INFO: Pod "pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64" satisfied condition "Succeeded or Failed" +Feb 25 09:47:07.020: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64 container agnhost-container: +STEP: delete the pod 02/25/23 09:47:07.035 +Feb 25 09:47:07.054: INFO: Waiting for pod pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64 to disappear +Feb 25 09:47:07.060: INFO: Pod pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64 no longer exists +[AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 -Dec 14 15:36:52.778: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-lifecycle-hook-4607" for this suite. 12/14/22 15:36:52.784 -{"msg":"PASSED [sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance]","completed":2,"skipped":13,"failed":0} +Feb 25 09:47:07.061: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-1842" for this suite. 02/25/23 09:47:07.067 +{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","completed":2,"skipped":55,"failed":0} ------------------------------ -• [SLOW TEST] [8.184 seconds] -[sig-node] Container Lifecycle Hook -test/e2e/common/node/framework.go:23 - when create a pod with lifecycle hook - test/e2e/common/node/lifecycle_hook.go:46 - should execute poststart exec hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:97 +• [4.147 seconds] +[sig-storage] Projected configMap +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:98 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Container Lifecycle Hook + [BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:36:44.632 - Dec 14 15:36:44.632: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-lifecycle-hook 12/14/22 15:36:44.636 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:36:44.661 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:36:44.665 - [BeforeEach] when create a pod with lifecycle hook - test/e2e/common/node/lifecycle_hook.go:55 - STEP: create the container to handle the HTTPGet hook request. 12/14/22 15:36:44.676 - Dec 14 15:36:44.687: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-4607" to be "running and ready" - Dec 14 15:36:44.692: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 4.595169ms - Dec 14 15:36:44.692: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) - Dec 14 15:36:46.704: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 2.016801768s - Dec 14 15:36:46.704: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) - Dec 14 15:36:46.704: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" - [It] should execute poststart exec hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:97 - STEP: create the pod with lifecycle hook 12/14/22 15:36:46.711 - Dec 14 15:36:46.720: INFO: Waiting up to 5m0s for pod "pod-with-poststart-exec-hook" in namespace "container-lifecycle-hook-4607" to be "running and ready" - Dec 14 15:36:46.729: INFO: Pod "pod-with-poststart-exec-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 8.514546ms - Dec 14 15:36:46.729: INFO: The phase of Pod pod-with-poststart-exec-hook is Pending, waiting for it to be Running (with Ready = true) - Dec 14 15:36:48.737: INFO: Pod "pod-with-poststart-exec-hook": Phase="Running", Reason="", readiness=true. Elapsed: 2.016717908s - Dec 14 15:36:48.737: INFO: The phase of Pod pod-with-poststart-exec-hook is Running (Ready = true) - Dec 14 15:36:48.737: INFO: Pod "pod-with-poststart-exec-hook" satisfied condition "running and ready" - STEP: check poststart hook 12/14/22 15:36:48.741 - STEP: delete the pod with lifecycle hook 12/14/22 15:36:48.751 - Dec 14 15:36:48.759: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear - Dec 14 15:36:48.769: INFO: Pod pod-with-poststart-exec-hook still exists - Dec 14 15:36:50.770: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear - Dec 14 15:36:50.783: INFO: Pod pod-with-poststart-exec-hook still exists - Dec 14 15:36:52.770: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear - Dec 14 15:36:52.777: INFO: Pod pod-with-poststart-exec-hook no longer exists - [AfterEach] [sig-node] Container Lifecycle Hook + STEP: Creating a kubernetes client 02/25/23 09:47:02.933 + Feb 25 09:47:02.934: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 09:47:02.936 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:02.963 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:02.968 + [It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:98 + STEP: Creating configMap with name projected-configmap-test-volume-map-80010730-b6b1-44d8-9efc-35c9ef6ee5ff 02/25/23 09:47:02.974 + STEP: Creating a pod to test consume configMaps 02/25/23 09:47:02.982 + Feb 25 09:47:02.997: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64" in namespace "projected-1842" to be "Succeeded or Failed" + Feb 25 09:47:03.006: INFO: Pod "pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64": Phase="Pending", Reason="", readiness=false. Elapsed: 8.95658ms + Feb 25 09:47:05.034: INFO: Pod "pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64": Phase="Pending", Reason="", readiness=false. Elapsed: 2.037508093s + Feb 25 09:47:07.014: INFO: Pod "pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.0174184s + STEP: Saw pod success 02/25/23 09:47:07.014 + Feb 25 09:47:07.015: INFO: Pod "pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64" satisfied condition "Succeeded or Failed" + Feb 25 09:47:07.020: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64 container agnhost-container: + STEP: delete the pod 02/25/23 09:47:07.035 + Feb 25 09:47:07.054: INFO: Waiting for pod pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64 to disappear + Feb 25 09:47:07.060: INFO: Pod pod-projected-configmaps-fc1e5490-c923-49be-bf2f-eeaeee7a8f64 no longer exists + [AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 - Dec 14 15:36:52.778: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-lifecycle-hook-4607" for this suite. 12/14/22 15:36:52.784 + Feb 25 09:47:07.061: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-1842" for this suite. 02/25/23 09:47:07.067 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSS ------------------------------ -[sig-node] Containers - should be able to override the image's default command (container entrypoint) [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:72 -[BeforeEach] [sig-node] Containers +[sig-network] DNS + should provide /etc/hosts entries for the cluster [Conformance] + test/e2e/network/dns.go:117 +[BeforeEach] [sig-network] DNS test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:36:52.817 -Dec 14 15:36:52.817: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename containers 12/14/22 15:36:52.821 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:36:52.849 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:36:52.853 -[It] should be able to override the image's default command (container entrypoint) [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:72 -STEP: Creating a pod to test override command 12/14/22 15:36:52.857 -Dec 14 15:36:52.871: INFO: Waiting up to 5m0s for pod "client-containers-c7923adf-3554-4654-a19f-9435cc30915e" in namespace "containers-6871" to be "Succeeded or Failed" -Dec 14 15:36:52.875: INFO: Pod "client-containers-c7923adf-3554-4654-a19f-9435cc30915e": Phase="Pending", Reason="", readiness=false. Elapsed: 4.20706ms -Dec 14 15:36:54.884: INFO: Pod "client-containers-c7923adf-3554-4654-a19f-9435cc30915e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012932343s -Dec 14 15:36:56.880: INFO: Pod "client-containers-c7923adf-3554-4654-a19f-9435cc30915e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.009760051s -STEP: Saw pod success 12/14/22 15:36:56.88 -Dec 14 15:36:56.881: INFO: Pod "client-containers-c7923adf-3554-4654-a19f-9435cc30915e" satisfied condition "Succeeded or Failed" -Dec 14 15:36:56.886: INFO: Trying to get logs from node kiedai4iakae-3 pod client-containers-c7923adf-3554-4654-a19f-9435cc30915e container agnhost-container: -STEP: delete the pod 12/14/22 15:36:56.897 -Dec 14 15:36:56.913: INFO: Waiting for pod client-containers-c7923adf-3554-4654-a19f-9435cc30915e to disappear -Dec 14 15:36:56.918: INFO: Pod client-containers-c7923adf-3554-4654-a19f-9435cc30915e no longer exists -[AfterEach] [sig-node] Containers +STEP: Creating a kubernetes client 02/25/23 09:47:07.082 +Feb 25 09:47:07.082: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename dns 02/25/23 09:47:07.087 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:07.114 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:07.118 +[It] should provide /etc/hosts entries for the cluster [Conformance] + test/e2e/network/dns.go:117 +STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-3399.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-3399.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;sleep 1; done + 02/25/23 09:47:07.122 +STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-3399.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-3399.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;sleep 1; done + 02/25/23 09:47:07.122 +STEP: creating a pod to probe /etc/hosts 02/25/23 09:47:07.122 +STEP: submitting the pod to kubernetes 02/25/23 09:47:07.122 +Feb 25 09:47:07.139: INFO: Waiting up to 15m0s for pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3" in namespace "dns-3399" to be "running" +Feb 25 09:47:07.145: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 5.587129ms +Feb 25 09:47:09.160: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01975396s +Feb 25 09:47:11.154: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 4.014550457s +Feb 25 09:47:13.168: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 6.028159037s +Feb 25 09:47:15.155: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 8.015039902s +Feb 25 09:47:17.153: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 10.013123334s +Feb 25 09:47:19.154: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 12.01462681s +Feb 25 09:47:21.154: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 14.014449655s +Feb 25 09:47:23.152: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 16.01262724s +Feb 25 09:47:25.154: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 18.014429899s +Feb 25 09:47:27.160: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 20.01984133s +Feb 25 09:47:29.152: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Running", Reason="", readiness=true. Elapsed: 22.012316079s +Feb 25 09:47:29.152: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3" satisfied condition "running" +STEP: retrieving the pod 02/25/23 09:47:29.152 +STEP: looking for the results for each expected name from probers 02/25/23 09:47:29.16 +Feb 25 09:47:29.196: INFO: DNS probes using dns-3399/dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3 succeeded + +STEP: deleting the pod 02/25/23 09:47:29.196 +[AfterEach] [sig-network] DNS test/e2e/framework/framework.go:187 -Dec 14 15:36:56.918: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "containers-6871" for this suite. 12/14/22 15:36:56.924 -{"msg":"PASSED [sig-node] Containers should be able to override the image's default command (container entrypoint) [NodeConformance] [Conformance]","completed":3,"skipped":16,"failed":0} +Feb 25 09:47:29.220: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "dns-3399" for this suite. 02/25/23 09:47:29.227 +{"msg":"PASSED [sig-network] DNS should provide /etc/hosts entries for the cluster [Conformance]","completed":3,"skipped":61,"failed":0} ------------------------------ -• [4.115 seconds] -[sig-node] Containers -test/e2e/common/node/framework.go:23 - should be able to override the image's default command (container entrypoint) [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:72 +• [SLOW TEST] [22.158 seconds] +[sig-network] DNS +test/e2e/network/common/framework.go:23 + should provide /etc/hosts entries for the cluster [Conformance] + test/e2e/network/dns.go:117 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Containers + [BeforeEach] [sig-network] DNS test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:36:52.817 - Dec 14 15:36:52.817: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename containers 12/14/22 15:36:52.821 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:36:52.849 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:36:52.853 - [It] should be able to override the image's default command (container entrypoint) [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:72 - STEP: Creating a pod to test override command 12/14/22 15:36:52.857 - Dec 14 15:36:52.871: INFO: Waiting up to 5m0s for pod "client-containers-c7923adf-3554-4654-a19f-9435cc30915e" in namespace "containers-6871" to be "Succeeded or Failed" - Dec 14 15:36:52.875: INFO: Pod "client-containers-c7923adf-3554-4654-a19f-9435cc30915e": Phase="Pending", Reason="", readiness=false. Elapsed: 4.20706ms - Dec 14 15:36:54.884: INFO: Pod "client-containers-c7923adf-3554-4654-a19f-9435cc30915e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012932343s - Dec 14 15:36:56.880: INFO: Pod "client-containers-c7923adf-3554-4654-a19f-9435cc30915e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.009760051s - STEP: Saw pod success 12/14/22 15:36:56.88 - Dec 14 15:36:56.881: INFO: Pod "client-containers-c7923adf-3554-4654-a19f-9435cc30915e" satisfied condition "Succeeded or Failed" - Dec 14 15:36:56.886: INFO: Trying to get logs from node kiedai4iakae-3 pod client-containers-c7923adf-3554-4654-a19f-9435cc30915e container agnhost-container: - STEP: delete the pod 12/14/22 15:36:56.897 - Dec 14 15:36:56.913: INFO: Waiting for pod client-containers-c7923adf-3554-4654-a19f-9435cc30915e to disappear - Dec 14 15:36:56.918: INFO: Pod client-containers-c7923adf-3554-4654-a19f-9435cc30915e no longer exists - [AfterEach] [sig-node] Containers + STEP: Creating a kubernetes client 02/25/23 09:47:07.082 + Feb 25 09:47:07.082: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename dns 02/25/23 09:47:07.087 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:07.114 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:07.118 + [It] should provide /etc/hosts entries for the cluster [Conformance] + test/e2e/network/dns.go:117 + STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-3399.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-3399.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;sleep 1; done + 02/25/23 09:47:07.122 + STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-3399.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-3399.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;sleep 1; done + 02/25/23 09:47:07.122 + STEP: creating a pod to probe /etc/hosts 02/25/23 09:47:07.122 + STEP: submitting the pod to kubernetes 02/25/23 09:47:07.122 + Feb 25 09:47:07.139: INFO: Waiting up to 15m0s for pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3" in namespace "dns-3399" to be "running" + Feb 25 09:47:07.145: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 5.587129ms + Feb 25 09:47:09.160: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01975396s + Feb 25 09:47:11.154: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 4.014550457s + Feb 25 09:47:13.168: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 6.028159037s + Feb 25 09:47:15.155: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 8.015039902s + Feb 25 09:47:17.153: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 10.013123334s + Feb 25 09:47:19.154: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 12.01462681s + Feb 25 09:47:21.154: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 14.014449655s + Feb 25 09:47:23.152: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 16.01262724s + Feb 25 09:47:25.154: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 18.014429899s + Feb 25 09:47:27.160: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Pending", Reason="", readiness=false. Elapsed: 20.01984133s + Feb 25 09:47:29.152: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3": Phase="Running", Reason="", readiness=true. Elapsed: 22.012316079s + Feb 25 09:47:29.152: INFO: Pod "dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3" satisfied condition "running" + STEP: retrieving the pod 02/25/23 09:47:29.152 + STEP: looking for the results for each expected name from probers 02/25/23 09:47:29.16 + Feb 25 09:47:29.196: INFO: DNS probes using dns-3399/dns-test-0231e91d-8365-4cac-ac74-602b0c0246a3 succeeded + + STEP: deleting the pod 02/25/23 09:47:29.196 + [AfterEach] [sig-network] DNS test/e2e/framework/framework.go:187 - Dec 14 15:36:56.918: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "containers-6871" for this suite. 12/14/22 15:36:56.924 + Feb 25 09:47:29.220: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "dns-3399" for this suite. 02/25/23 09:47:29.227 << End Captured GinkgoWriter Output ------------------------------ -[sig-storage] Projected downwardAPI - should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:248 -[BeforeEach] [sig-storage] Projected downwardAPI +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-network] Proxy version v1 + should proxy through a service and a pod [Conformance] + test/e2e/network/proxy.go:101 +[BeforeEach] version v1 test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:36:56.934 -Dec 14 15:36:56.935: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 15:36:56.94 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:36:56.966 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:36:56.971 -[BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 -[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:248 -STEP: Creating a pod to test downward API volume plugin 12/14/22 15:36:56.975 -Dec 14 15:36:56.990: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389" in namespace "projected-9393" to be "Succeeded or Failed" -Dec 14 15:36:56.994: INFO: Pod "downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389": Phase="Pending", Reason="", readiness=false. Elapsed: 3.847335ms -Dec 14 15:36:59.004: INFO: Pod "downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013707533s -Dec 14 15:37:01.004: INFO: Pod "downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013650558s -STEP: Saw pod success 12/14/22 15:37:01.004 -Dec 14 15:37:01.005: INFO: Pod "downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389" satisfied condition "Succeeded or Failed" -Dec 14 15:37:01.011: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389 container client-container: -STEP: delete the pod 12/14/22 15:37:01.023 -Dec 14 15:37:01.048: INFO: Waiting for pod downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389 to disappear -Dec 14 15:37:01.060: INFO: Pod downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389 no longer exists -[AfterEach] [sig-storage] Projected downwardAPI +STEP: Creating a kubernetes client 02/25/23 09:47:29.244 +Feb 25 09:47:29.245: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename proxy 02/25/23 09:47:29.247 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:29.275 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:29.281 +[It] should proxy through a service and a pod [Conformance] + test/e2e/network/proxy.go:101 +STEP: starting an echo server on multiple ports 02/25/23 09:47:29.316 +STEP: creating replication controller proxy-service-b9nvz in namespace proxy-7283 02/25/23 09:47:29.317 +I0225 09:47:29.346611 13 runners.go:193] Created replication controller with name: proxy-service-b9nvz, namespace: proxy-7283, replica count: 1 +I0225 09:47:30.400096 13 runners.go:193] proxy-service-b9nvz Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0225 09:47:31.401256 13 runners.go:193] proxy-service-b9nvz Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Feb 25 09:47:31.414: INFO: setup took 2.126327178s, starting test cases +STEP: running 16 cases, 20 attempts per case, 320 total attempts 02/25/23 09:47:31.415 +Feb 25 09:47:31.447: INFO: (0) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 32.141619ms) +Feb 25 09:47:31.447: INFO: (0) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 32.100638ms) +Feb 25 09:47:31.448: INFO: (0) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 32.453503ms) +Feb 25 09:47:31.448: INFO: (0) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 32.671455ms) +Feb 25 09:47:31.452: INFO: (0) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 36.243746ms) +Feb 25 09:47:31.452: INFO: (0) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 36.797317ms) +Feb 25 09:47:31.452: INFO: (0) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 36.858771ms) +Feb 25 09:47:31.462: INFO: (0) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 46.755964ms) +Feb 25 09:47:31.462: INFO: (0) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 46.583959ms) +Feb 25 09:47:31.470: INFO: (0) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 54.372461ms) +Feb 25 09:47:31.471: INFO: (0) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 55.701103ms) +Feb 25 09:47:31.471: INFO: (0) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 55.967265ms) +Feb 25 09:47:31.474: INFO: (0) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 59.458794ms) +Feb 25 09:47:31.474: INFO: (0) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 58.568191ms) +Feb 25 09:47:31.474: INFO: (0) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 58.960303ms) +Feb 25 09:47:31.477: INFO: (0) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 15.295083ms) +Feb 25 09:47:31.493: INFO: (1) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 15.051421ms) +Feb 25 09:47:31.495: INFO: (1) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: ... (200; 19.739835ms) +Feb 25 09:47:31.496: INFO: (1) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 19.05938ms) +Feb 25 09:47:31.507: INFO: (1) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 30.198088ms) +Feb 25 09:47:31.507: INFO: (1) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 29.501172ms) +Feb 25 09:47:31.507: INFO: (1) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 29.669712ms) +Feb 25 09:47:31.507: INFO: (1) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 29.849818ms) +Feb 25 09:47:31.508: INFO: (1) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 30.994413ms) +Feb 25 09:47:31.508: INFO: (1) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 29.624896ms) +Feb 25 09:47:31.508: INFO: (1) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 31.458797ms) +Feb 25 09:47:31.511: INFO: (1) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 33.89476ms) +Feb 25 09:47:31.523: INFO: (2) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 12.348071ms) +Feb 25 09:47:31.526: INFO: (2) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 14.899425ms) +Feb 25 09:47:31.526: INFO: (2) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 15.116195ms) +Feb 25 09:47:31.532: INFO: (2) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 20.411638ms) +Feb 25 09:47:31.532: INFO: (2) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 20.883172ms) +Feb 25 09:47:31.532: INFO: (2) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 25.241926ms) +Feb 25 09:47:31.537: INFO: (2) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 25.46251ms) +Feb 25 09:47:31.540: INFO: (2) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 28.981764ms) +Feb 25 09:47:31.541: INFO: (2) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 29.278048ms) +Feb 25 09:47:31.541: INFO: (2) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 29.576339ms) +Feb 25 09:47:31.541: INFO: (2) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 30.043327ms) +Feb 25 09:47:31.541: INFO: (2) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 29.712769ms) +Feb 25 09:47:31.541: INFO: (2) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 30.004559ms) +Feb 25 09:47:31.541: INFO: (2) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 30.190181ms) +Feb 25 09:47:31.561: INFO: (3) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 19.048686ms) +Feb 25 09:47:31.561: INFO: (3) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 19.305864ms) +Feb 25 09:47:31.562: INFO: (3) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 30.440911ms) +Feb 25 09:47:31.572: INFO: (3) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 30.590457ms) +Feb 25 09:47:31.578: INFO: (3) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 36.335843ms) +Feb 25 09:47:31.579: INFO: (3) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 36.670174ms) +Feb 25 09:47:31.579: INFO: (3) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 37.01753ms) +Feb 25 09:47:31.586: INFO: (3) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 43.955258ms) +Feb 25 09:47:31.586: INFO: (3) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 44.295918ms) +Feb 25 09:47:31.587: INFO: (3) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 44.830985ms) +Feb 25 09:47:31.587: INFO: (3) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 44.214664ms) +Feb 25 09:47:31.588: INFO: (3) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 45.643673ms) +Feb 25 09:47:31.599: INFO: (4) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 10.628941ms) +Feb 25 09:47:31.603: INFO: (4) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 14.269641ms) +Feb 25 09:47:31.619: INFO: (4) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 30.63984ms) +Feb 25 09:47:31.619: INFO: (4) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 30.902222ms) +Feb 25 09:47:31.620: INFO: (4) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 34.218203ms) +Feb 25 09:47:31.626: INFO: (4) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 36.534126ms) +Feb 25 09:47:31.626: INFO: (4) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 36.681477ms) +Feb 25 09:47:31.626: INFO: (4) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 36.94502ms) +Feb 25 09:47:31.627: INFO: (4) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 38.500696ms) +Feb 25 09:47:31.631: INFO: (4) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 42.162516ms) +Feb 25 09:47:31.638: INFO: (4) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 49.302649ms) +Feb 25 09:47:31.646: INFO: (4) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 58.137285ms) +Feb 25 09:47:31.647: INFO: (4) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 58.154953ms) +Feb 25 09:47:31.647: INFO: (4) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 58.396006ms) +Feb 25 09:47:31.674: INFO: (5) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 23.961039ms) +Feb 25 09:47:31.674: INFO: (5) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 26.250566ms) +Feb 25 09:47:31.675: INFO: (5) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 27.192037ms) +Feb 25 09:47:31.675: INFO: (5) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 27.440696ms) +Feb 25 09:47:31.677: INFO: (5) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 27.305086ms) +Feb 25 09:47:31.681: INFO: (5) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 30.758905ms) +Feb 25 09:47:31.681: INFO: (5) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 34.197327ms) +Feb 25 09:47:31.682: INFO: (5) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 34.493862ms) +Feb 25 09:47:31.682: INFO: (5) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 32.933033ms) +Feb 25 09:47:31.682: INFO: (5) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 33.549599ms) +Feb 25 09:47:31.685: INFO: (5) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 37.693475ms) +Feb 25 09:47:31.688: INFO: (5) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 38.753856ms) +Feb 25 09:47:31.703: INFO: (6) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 12.337153ms) +Feb 25 09:47:31.709: INFO: (6) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 17.798943ms) +Feb 25 09:47:31.714: INFO: (6) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 25.11267ms) +Feb 25 09:47:31.715: INFO: (6) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 25.814807ms) +Feb 25 09:47:31.715: INFO: (6) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 25.720985ms) +Feb 25 09:47:31.720: INFO: (6) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 27.846034ms) +Feb 25 09:47:31.720: INFO: (6) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 25.602538ms) +Feb 25 09:47:31.720: INFO: (6) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 29.217914ms) +Feb 25 09:47:31.722: INFO: (6) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 27.437738ms) +Feb 25 09:47:31.722: INFO: (6) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 27.929871ms) +Feb 25 09:47:31.722: INFO: (6) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 33.645243ms) +Feb 25 09:47:31.723: INFO: (6) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 32.563487ms) +Feb 25 09:47:31.724: INFO: (6) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test (200; 28.858439ms) +Feb 25 09:47:31.725: INFO: (6) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 32.925101ms) +Feb 25 09:47:31.726: INFO: (6) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 30.693304ms) +Feb 25 09:47:31.741: INFO: (7) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 14.854337ms) +Feb 25 09:47:31.743: INFO: (7) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 16.910301ms) +Feb 25 09:47:31.744: INFO: (7) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 16.386957ms) +Feb 25 09:47:31.745: INFO: (7) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 18.749455ms) +Feb 25 09:47:31.748: INFO: (7) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 20.512053ms) +Feb 25 09:47:31.749: INFO: (7) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 19.163812ms) +Feb 25 09:47:31.749: INFO: (7) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 20.345691ms) +Feb 25 09:47:31.749: INFO: (7) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 22.166928ms) +Feb 25 09:47:31.749: INFO: (7) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 20.864658ms) +Feb 25 09:47:31.752: INFO: (7) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 23.842479ms) +Feb 25 09:47:31.753: INFO: (7) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 25.130162ms) +Feb 25 09:47:31.753: INFO: (7) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 24.159085ms) +Feb 25 09:47:31.754: INFO: (7) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 25.117235ms) +Feb 25 09:47:31.757: INFO: (7) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 27.921465ms) +Feb 25 09:47:31.762: INFO: (8) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 5.00714ms) +Feb 25 09:47:31.769: INFO: (8) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 11.451659ms) +Feb 25 09:47:31.771: INFO: (8) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 13.566274ms) +Feb 25 09:47:31.772: INFO: (8) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 13.583961ms) +Feb 25 09:47:31.775: INFO: (8) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 16.310134ms) +Feb 25 09:47:31.776: INFO: (8) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 17.386966ms) +Feb 25 09:47:31.776: INFO: (8) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 17.419673ms) +Feb 25 09:47:31.779: INFO: (8) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 21.913776ms) +Feb 25 09:47:31.779: INFO: (8) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 20.754271ms) +Feb 25 09:47:31.780: INFO: (8) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 22.366004ms) +Feb 25 09:47:31.781: INFO: (8) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 23.141403ms) +Feb 25 09:47:31.783: INFO: (8) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 24.537397ms) +Feb 25 09:47:31.783: INFO: (8) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 24.841813ms) +Feb 25 09:47:31.783: INFO: (8) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 24.606204ms) +Feb 25 09:47:31.784: INFO: (8) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 25.541075ms) +Feb 25 09:47:31.784: INFO: (8) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 24.640054ms) +Feb 25 09:47:31.811: INFO: (9) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 26.315009ms) +Feb 25 09:47:31.816: INFO: (9) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 31.977128ms) +Feb 25 09:47:31.817: INFO: (9) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 32.334044ms) +Feb 25 09:47:31.817: INFO: (9) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 31.530899ms) +Feb 25 09:47:31.817: INFO: (9) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 32.091813ms) +Feb 25 09:47:31.817: INFO: (9) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 32.585ms) +Feb 25 09:47:31.817: INFO: (9) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 31.865884ms) +Feb 25 09:47:31.818: INFO: (9) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 32.819845ms) +Feb 25 09:47:31.819: INFO: (9) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 33.768066ms) +Feb 25 09:47:31.833: INFO: (10) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 14.181653ms) +Feb 25 09:47:31.836: INFO: (10) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 17.148983ms) +Feb 25 09:47:31.837: INFO: (10) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 17.668687ms) +Feb 25 09:47:31.839: INFO: (10) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 19.727878ms) +Feb 25 09:47:31.839: INFO: (10) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: ... (200; 25.605966ms) +Feb 25 09:47:31.848: INFO: (10) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 28.468715ms) +Feb 25 09:47:31.863: INFO: (11) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 15.47174ms) +Feb 25 09:47:31.869: INFO: (11) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 20.008589ms) +Feb 25 09:47:31.869: INFO: (11) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 20.561854ms) +Feb 25 09:47:31.871: INFO: (11) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 22.07716ms) +Feb 25 09:47:31.874: INFO: (11) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 24.991596ms) +Feb 25 09:47:31.877: INFO: (11) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 27.569835ms) +Feb 25 09:47:31.877: INFO: (11) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 30.53559ms) +Feb 25 09:47:31.879: INFO: (11) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 30.143855ms) +Feb 25 09:47:31.880: INFO: (11) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 31.343715ms) +Feb 25 09:47:31.880: INFO: (11) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 30.65419ms) +Feb 25 09:47:31.880: INFO: (11) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 31.110562ms) +Feb 25 09:47:31.891: INFO: (12) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 10.060171ms) +Feb 25 09:47:31.893: INFO: (12) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 11.06793ms) +Feb 25 09:47:31.895: INFO: (12) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 12.093136ms) +Feb 25 09:47:31.902: INFO: (12) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 18.46698ms) +Feb 25 09:47:31.902: INFO: (12) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 18.918072ms) +Feb 25 09:47:31.902: INFO: (12) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 19.019598ms) +Feb 25 09:47:31.903: INFO: (12) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 22.400326ms) +Feb 25 09:47:31.904: INFO: (12) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 22.010683ms) +Feb 25 09:47:31.905: INFO: (12) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 11.710871ms) +Feb 25 09:47:31.925: INFO: (13) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 12.352945ms) +Feb 25 09:47:31.932: INFO: (13) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 19.108527ms) +Feb 25 09:47:31.935: INFO: (13) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: ... (200; 30.856191ms) +Feb 25 09:47:31.944: INFO: (13) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 30.927994ms) +Feb 25 09:47:31.944: INFO: (13) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 31.580517ms) +Feb 25 09:47:31.946: INFO: (13) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 33.110299ms) +Feb 25 09:47:31.946: INFO: (13) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 33.050115ms) +Feb 25 09:47:31.961: INFO: (14) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 14.272457ms) +Feb 25 09:47:31.961: INFO: (14) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 14.482212ms) +Feb 25 09:47:31.961: INFO: (14) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 14.365683ms) +Feb 25 09:47:31.967: INFO: (14) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 20.191577ms) +Feb 25 09:47:31.972: INFO: (14) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 24.818068ms) +Feb 25 09:47:31.972: INFO: (14) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 25.06564ms) +Feb 25 09:47:31.972: INFO: (14) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 24.624611ms) +Feb 25 09:47:31.973: INFO: (14) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 25.393557ms) +Feb 25 09:47:31.973: INFO: (14) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 24.668536ms) +Feb 25 09:47:31.974: INFO: (14) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 26.37823ms) +Feb 25 09:47:31.974: INFO: (14) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 26.182178ms) +Feb 25 09:47:31.974: INFO: (14) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 26.120174ms) +Feb 25 09:47:31.978: INFO: (14) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 30.044532ms) +Feb 25 09:47:31.978: INFO: (14) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 30.240967ms) +Feb 25 09:47:31.980: INFO: (14) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test (200; 26.439076ms) +Feb 25 09:47:32.007: INFO: (15) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 25.819238ms) +Feb 25 09:47:32.007: INFO: (15) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 26.226269ms) +Feb 25 09:47:32.007: INFO: (15) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 26.363883ms) +Feb 25 09:47:32.007: INFO: (15) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 25.092156ms) +Feb 25 09:47:32.009: INFO: (15) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 11.906952ms) +Feb 25 09:47:32.025: INFO: (16) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 12.012694ms) +Feb 25 09:47:32.027: INFO: (16) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 13.017251ms) +Feb 25 09:47:32.027: INFO: (16) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: ... (200; 21.262245ms) +Feb 25 09:47:32.035: INFO: (16) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 21.465893ms) +Feb 25 09:47:32.036: INFO: (16) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 21.813064ms) +Feb 25 09:47:32.036: INFO: (16) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 21.756711ms) +Feb 25 09:47:32.037: INFO: (16) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 23.471012ms) +Feb 25 09:47:32.037: INFO: (16) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 23.209219ms) +Feb 25 09:47:32.038: INFO: (16) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 23.625235ms) +Feb 25 09:47:32.038: INFO: (16) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 24.498494ms) +Feb 25 09:47:32.040: INFO: (16) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 25.965387ms) +Feb 25 09:47:32.062: INFO: (17) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 21.538867ms) +Feb 25 09:47:32.064: INFO: (17) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 21.903374ms) +Feb 25 09:47:32.064: INFO: (17) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 23.440285ms) +Feb 25 09:47:32.067: INFO: (17) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 24.517662ms) +Feb 25 09:47:32.067: INFO: (17) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 25.886767ms) +Feb 25 09:47:32.067: INFO: (17) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test (200; 26.362812ms) +Feb 25 09:47:32.068: INFO: (17) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 26.605805ms) +Feb 25 09:47:32.068: INFO: (17) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 27.444871ms) +Feb 25 09:47:32.068: INFO: (17) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 26.68948ms) +Feb 25 09:47:32.069: INFO: (17) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 28.713331ms) +Feb 25 09:47:32.070: INFO: (17) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 27.519819ms) +Feb 25 09:47:32.070: INFO: (17) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 28.952531ms) +Feb 25 09:47:32.070: INFO: (17) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 29.084565ms) +Feb 25 09:47:32.082: INFO: (18) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 10.746554ms) +Feb 25 09:47:32.083: INFO: (18) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 12.234611ms) +Feb 25 09:47:32.084: INFO: (18) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 13.399517ms) +Feb 25 09:47:32.085: INFO: (18) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 13.54368ms) +Feb 25 09:47:32.086: INFO: (18) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 14.701604ms) +Feb 25 09:47:32.087: INFO: (18) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 16.017447ms) +Feb 25 09:47:32.090: INFO: (18) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 18.606878ms) +Feb 25 09:47:32.090: INFO: (18) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 27.632653ms) +Feb 25 09:47:32.100: INFO: (18) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 27.978582ms) +Feb 25 09:47:32.109: INFO: (19) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 8.02944ms) +Feb 25 09:47:32.110: INFO: (19) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 9.11425ms) +Feb 25 09:47:32.121: INFO: (19) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 18.80755ms) +Feb 25 09:47:32.121: INFO: (19) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 19.151867ms) +Feb 25 09:47:32.121: INFO: (19) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 18.797144ms) +Feb 25 09:47:32.122: INFO: (19) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: > - [BeforeEach] [sig-storage] Projected downwardAPI + [BeforeEach] version v1 test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:36:56.934 - Dec 14 15:36:56.935: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 15:36:56.94 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:36:56.966 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:36:56.971 - [BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 - [It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:248 - STEP: Creating a pod to test downward API volume plugin 12/14/22 15:36:56.975 - Dec 14 15:36:56.990: INFO: Waiting up to 5m0s for pod "downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389" in namespace "projected-9393" to be "Succeeded or Failed" - Dec 14 15:36:56.994: INFO: Pod "downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389": Phase="Pending", Reason="", readiness=false. Elapsed: 3.847335ms - Dec 14 15:36:59.004: INFO: Pod "downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013707533s - Dec 14 15:37:01.004: INFO: Pod "downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013650558s - STEP: Saw pod success 12/14/22 15:37:01.004 - Dec 14 15:37:01.005: INFO: Pod "downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389" satisfied condition "Succeeded or Failed" - Dec 14 15:37:01.011: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389 container client-container: - STEP: delete the pod 12/14/22 15:37:01.023 - Dec 14 15:37:01.048: INFO: Waiting for pod downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389 to disappear - Dec 14 15:37:01.060: INFO: Pod downwardapi-volume-e6e83476-7047-47e8-b52f-d3a474c4e389 no longer exists - [AfterEach] [sig-storage] Projected downwardAPI + STEP: Creating a kubernetes client 02/25/23 09:47:29.244 + Feb 25 09:47:29.245: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename proxy 02/25/23 09:47:29.247 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:29.275 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:29.281 + [It] should proxy through a service and a pod [Conformance] + test/e2e/network/proxy.go:101 + STEP: starting an echo server on multiple ports 02/25/23 09:47:29.316 + STEP: creating replication controller proxy-service-b9nvz in namespace proxy-7283 02/25/23 09:47:29.317 + I0225 09:47:29.346611 13 runners.go:193] Created replication controller with name: proxy-service-b9nvz, namespace: proxy-7283, replica count: 1 + I0225 09:47:30.400096 13 runners.go:193] proxy-service-b9nvz Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + I0225 09:47:31.401256 13 runners.go:193] proxy-service-b9nvz Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + Feb 25 09:47:31.414: INFO: setup took 2.126327178s, starting test cases + STEP: running 16 cases, 20 attempts per case, 320 total attempts 02/25/23 09:47:31.415 + Feb 25 09:47:31.447: INFO: (0) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 32.141619ms) + Feb 25 09:47:31.447: INFO: (0) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 32.100638ms) + Feb 25 09:47:31.448: INFO: (0) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 32.453503ms) + Feb 25 09:47:31.448: INFO: (0) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 32.671455ms) + Feb 25 09:47:31.452: INFO: (0) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 36.243746ms) + Feb 25 09:47:31.452: INFO: (0) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 36.797317ms) + Feb 25 09:47:31.452: INFO: (0) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 36.858771ms) + Feb 25 09:47:31.462: INFO: (0) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 46.755964ms) + Feb 25 09:47:31.462: INFO: (0) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 46.583959ms) + Feb 25 09:47:31.470: INFO: (0) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 54.372461ms) + Feb 25 09:47:31.471: INFO: (0) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 55.701103ms) + Feb 25 09:47:31.471: INFO: (0) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 55.967265ms) + Feb 25 09:47:31.474: INFO: (0) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 59.458794ms) + Feb 25 09:47:31.474: INFO: (0) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 58.568191ms) + Feb 25 09:47:31.474: INFO: (0) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 58.960303ms) + Feb 25 09:47:31.477: INFO: (0) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 15.295083ms) + Feb 25 09:47:31.493: INFO: (1) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 15.051421ms) + Feb 25 09:47:31.495: INFO: (1) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: ... (200; 19.739835ms) + Feb 25 09:47:31.496: INFO: (1) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 19.05938ms) + Feb 25 09:47:31.507: INFO: (1) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 30.198088ms) + Feb 25 09:47:31.507: INFO: (1) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 29.501172ms) + Feb 25 09:47:31.507: INFO: (1) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 29.669712ms) + Feb 25 09:47:31.507: INFO: (1) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 29.849818ms) + Feb 25 09:47:31.508: INFO: (1) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 30.994413ms) + Feb 25 09:47:31.508: INFO: (1) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 29.624896ms) + Feb 25 09:47:31.508: INFO: (1) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 31.458797ms) + Feb 25 09:47:31.511: INFO: (1) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 33.89476ms) + Feb 25 09:47:31.523: INFO: (2) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 12.348071ms) + Feb 25 09:47:31.526: INFO: (2) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 14.899425ms) + Feb 25 09:47:31.526: INFO: (2) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 15.116195ms) + Feb 25 09:47:31.532: INFO: (2) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 20.411638ms) + Feb 25 09:47:31.532: INFO: (2) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 20.883172ms) + Feb 25 09:47:31.532: INFO: (2) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 25.241926ms) + Feb 25 09:47:31.537: INFO: (2) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 25.46251ms) + Feb 25 09:47:31.540: INFO: (2) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 28.981764ms) + Feb 25 09:47:31.541: INFO: (2) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 29.278048ms) + Feb 25 09:47:31.541: INFO: (2) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 29.576339ms) + Feb 25 09:47:31.541: INFO: (2) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 30.043327ms) + Feb 25 09:47:31.541: INFO: (2) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 29.712769ms) + Feb 25 09:47:31.541: INFO: (2) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 30.004559ms) + Feb 25 09:47:31.541: INFO: (2) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 30.190181ms) + Feb 25 09:47:31.561: INFO: (3) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 19.048686ms) + Feb 25 09:47:31.561: INFO: (3) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 19.305864ms) + Feb 25 09:47:31.562: INFO: (3) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 30.440911ms) + Feb 25 09:47:31.572: INFO: (3) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 30.590457ms) + Feb 25 09:47:31.578: INFO: (3) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 36.335843ms) + Feb 25 09:47:31.579: INFO: (3) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 36.670174ms) + Feb 25 09:47:31.579: INFO: (3) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 37.01753ms) + Feb 25 09:47:31.586: INFO: (3) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 43.955258ms) + Feb 25 09:47:31.586: INFO: (3) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 44.295918ms) + Feb 25 09:47:31.587: INFO: (3) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 44.830985ms) + Feb 25 09:47:31.587: INFO: (3) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 44.214664ms) + Feb 25 09:47:31.588: INFO: (3) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 45.643673ms) + Feb 25 09:47:31.599: INFO: (4) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 10.628941ms) + Feb 25 09:47:31.603: INFO: (4) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 14.269641ms) + Feb 25 09:47:31.619: INFO: (4) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 30.63984ms) + Feb 25 09:47:31.619: INFO: (4) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 30.902222ms) + Feb 25 09:47:31.620: INFO: (4) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 34.218203ms) + Feb 25 09:47:31.626: INFO: (4) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 36.534126ms) + Feb 25 09:47:31.626: INFO: (4) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 36.681477ms) + Feb 25 09:47:31.626: INFO: (4) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 36.94502ms) + Feb 25 09:47:31.627: INFO: (4) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 38.500696ms) + Feb 25 09:47:31.631: INFO: (4) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 42.162516ms) + Feb 25 09:47:31.638: INFO: (4) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 49.302649ms) + Feb 25 09:47:31.646: INFO: (4) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 58.137285ms) + Feb 25 09:47:31.647: INFO: (4) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 58.154953ms) + Feb 25 09:47:31.647: INFO: (4) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 58.396006ms) + Feb 25 09:47:31.674: INFO: (5) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 23.961039ms) + Feb 25 09:47:31.674: INFO: (5) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 26.250566ms) + Feb 25 09:47:31.675: INFO: (5) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 27.192037ms) + Feb 25 09:47:31.675: INFO: (5) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 27.440696ms) + Feb 25 09:47:31.677: INFO: (5) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 27.305086ms) + Feb 25 09:47:31.681: INFO: (5) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 30.758905ms) + Feb 25 09:47:31.681: INFO: (5) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 34.197327ms) + Feb 25 09:47:31.682: INFO: (5) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 34.493862ms) + Feb 25 09:47:31.682: INFO: (5) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 32.933033ms) + Feb 25 09:47:31.682: INFO: (5) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 33.549599ms) + Feb 25 09:47:31.685: INFO: (5) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 37.693475ms) + Feb 25 09:47:31.688: INFO: (5) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 38.753856ms) + Feb 25 09:47:31.703: INFO: (6) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 12.337153ms) + Feb 25 09:47:31.709: INFO: (6) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 17.798943ms) + Feb 25 09:47:31.714: INFO: (6) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 25.11267ms) + Feb 25 09:47:31.715: INFO: (6) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 25.814807ms) + Feb 25 09:47:31.715: INFO: (6) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 25.720985ms) + Feb 25 09:47:31.720: INFO: (6) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 27.846034ms) + Feb 25 09:47:31.720: INFO: (6) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 25.602538ms) + Feb 25 09:47:31.720: INFO: (6) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 29.217914ms) + Feb 25 09:47:31.722: INFO: (6) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 27.437738ms) + Feb 25 09:47:31.722: INFO: (6) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 27.929871ms) + Feb 25 09:47:31.722: INFO: (6) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 33.645243ms) + Feb 25 09:47:31.723: INFO: (6) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 32.563487ms) + Feb 25 09:47:31.724: INFO: (6) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test (200; 28.858439ms) + Feb 25 09:47:31.725: INFO: (6) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 32.925101ms) + Feb 25 09:47:31.726: INFO: (6) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 30.693304ms) + Feb 25 09:47:31.741: INFO: (7) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 14.854337ms) + Feb 25 09:47:31.743: INFO: (7) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 16.910301ms) + Feb 25 09:47:31.744: INFO: (7) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 16.386957ms) + Feb 25 09:47:31.745: INFO: (7) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 18.749455ms) + Feb 25 09:47:31.748: INFO: (7) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 20.512053ms) + Feb 25 09:47:31.749: INFO: (7) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 19.163812ms) + Feb 25 09:47:31.749: INFO: (7) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 20.345691ms) + Feb 25 09:47:31.749: INFO: (7) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 22.166928ms) + Feb 25 09:47:31.749: INFO: (7) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 20.864658ms) + Feb 25 09:47:31.752: INFO: (7) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 23.842479ms) + Feb 25 09:47:31.753: INFO: (7) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 25.130162ms) + Feb 25 09:47:31.753: INFO: (7) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 24.159085ms) + Feb 25 09:47:31.754: INFO: (7) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 25.117235ms) + Feb 25 09:47:31.757: INFO: (7) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 27.921465ms) + Feb 25 09:47:31.762: INFO: (8) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 5.00714ms) + Feb 25 09:47:31.769: INFO: (8) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 11.451659ms) + Feb 25 09:47:31.771: INFO: (8) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 13.566274ms) + Feb 25 09:47:31.772: INFO: (8) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 13.583961ms) + Feb 25 09:47:31.775: INFO: (8) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 16.310134ms) + Feb 25 09:47:31.776: INFO: (8) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 17.386966ms) + Feb 25 09:47:31.776: INFO: (8) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 17.419673ms) + Feb 25 09:47:31.779: INFO: (8) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 21.913776ms) + Feb 25 09:47:31.779: INFO: (8) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 20.754271ms) + Feb 25 09:47:31.780: INFO: (8) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 22.366004ms) + Feb 25 09:47:31.781: INFO: (8) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 23.141403ms) + Feb 25 09:47:31.783: INFO: (8) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 24.537397ms) + Feb 25 09:47:31.783: INFO: (8) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 24.841813ms) + Feb 25 09:47:31.783: INFO: (8) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 24.606204ms) + Feb 25 09:47:31.784: INFO: (8) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 25.541075ms) + Feb 25 09:47:31.784: INFO: (8) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 24.640054ms) + Feb 25 09:47:31.811: INFO: (9) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 26.315009ms) + Feb 25 09:47:31.816: INFO: (9) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 31.977128ms) + Feb 25 09:47:31.817: INFO: (9) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 32.334044ms) + Feb 25 09:47:31.817: INFO: (9) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 31.530899ms) + Feb 25 09:47:31.817: INFO: (9) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 32.091813ms) + Feb 25 09:47:31.817: INFO: (9) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 32.585ms) + Feb 25 09:47:31.817: INFO: (9) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 31.865884ms) + Feb 25 09:47:31.818: INFO: (9) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 32.819845ms) + Feb 25 09:47:31.819: INFO: (9) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 33.768066ms) + Feb 25 09:47:31.833: INFO: (10) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 14.181653ms) + Feb 25 09:47:31.836: INFO: (10) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 17.148983ms) + Feb 25 09:47:31.837: INFO: (10) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 17.668687ms) + Feb 25 09:47:31.839: INFO: (10) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 19.727878ms) + Feb 25 09:47:31.839: INFO: (10) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: ... (200; 25.605966ms) + Feb 25 09:47:31.848: INFO: (10) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 28.468715ms) + Feb 25 09:47:31.863: INFO: (11) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 15.47174ms) + Feb 25 09:47:31.869: INFO: (11) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 20.008589ms) + Feb 25 09:47:31.869: INFO: (11) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 20.561854ms) + Feb 25 09:47:31.871: INFO: (11) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 22.07716ms) + Feb 25 09:47:31.874: INFO: (11) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 24.991596ms) + Feb 25 09:47:31.877: INFO: (11) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 27.569835ms) + Feb 25 09:47:31.877: INFO: (11) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 30.53559ms) + Feb 25 09:47:31.879: INFO: (11) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 30.143855ms) + Feb 25 09:47:31.880: INFO: (11) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 31.343715ms) + Feb 25 09:47:31.880: INFO: (11) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 30.65419ms) + Feb 25 09:47:31.880: INFO: (11) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 31.110562ms) + Feb 25 09:47:31.891: INFO: (12) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 10.060171ms) + Feb 25 09:47:31.893: INFO: (12) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 11.06793ms) + Feb 25 09:47:31.895: INFO: (12) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 12.093136ms) + Feb 25 09:47:31.902: INFO: (12) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 18.46698ms) + Feb 25 09:47:31.902: INFO: (12) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 18.918072ms) + Feb 25 09:47:31.902: INFO: (12) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 19.019598ms) + Feb 25 09:47:31.903: INFO: (12) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 22.400326ms) + Feb 25 09:47:31.904: INFO: (12) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 22.010683ms) + Feb 25 09:47:31.905: INFO: (12) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 11.710871ms) + Feb 25 09:47:31.925: INFO: (13) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 12.352945ms) + Feb 25 09:47:31.932: INFO: (13) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 19.108527ms) + Feb 25 09:47:31.935: INFO: (13) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: ... (200; 30.856191ms) + Feb 25 09:47:31.944: INFO: (13) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 30.927994ms) + Feb 25 09:47:31.944: INFO: (13) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 31.580517ms) + Feb 25 09:47:31.946: INFO: (13) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 33.110299ms) + Feb 25 09:47:31.946: INFO: (13) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 33.050115ms) + Feb 25 09:47:31.961: INFO: (14) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 14.272457ms) + Feb 25 09:47:31.961: INFO: (14) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 14.482212ms) + Feb 25 09:47:31.961: INFO: (14) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 14.365683ms) + Feb 25 09:47:31.967: INFO: (14) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 20.191577ms) + Feb 25 09:47:31.972: INFO: (14) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 24.818068ms) + Feb 25 09:47:31.972: INFO: (14) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 25.06564ms) + Feb 25 09:47:31.972: INFO: (14) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 24.624611ms) + Feb 25 09:47:31.973: INFO: (14) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 25.393557ms) + Feb 25 09:47:31.973: INFO: (14) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 24.668536ms) + Feb 25 09:47:31.974: INFO: (14) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 26.37823ms) + Feb 25 09:47:31.974: INFO: (14) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 26.182178ms) + Feb 25 09:47:31.974: INFO: (14) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 26.120174ms) + Feb 25 09:47:31.978: INFO: (14) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 30.044532ms) + Feb 25 09:47:31.978: INFO: (14) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 30.240967ms) + Feb 25 09:47:31.980: INFO: (14) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test (200; 26.439076ms) + Feb 25 09:47:32.007: INFO: (15) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 25.819238ms) + Feb 25 09:47:32.007: INFO: (15) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 26.226269ms) + Feb 25 09:47:32.007: INFO: (15) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 26.363883ms) + Feb 25 09:47:32.007: INFO: (15) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 25.092156ms) + Feb 25 09:47:32.009: INFO: (15) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 11.906952ms) + Feb 25 09:47:32.025: INFO: (16) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 12.012694ms) + Feb 25 09:47:32.027: INFO: (16) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 13.017251ms) + Feb 25 09:47:32.027: INFO: (16) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: ... (200; 21.262245ms) + Feb 25 09:47:32.035: INFO: (16) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 21.465893ms) + Feb 25 09:47:32.036: INFO: (16) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 21.813064ms) + Feb 25 09:47:32.036: INFO: (16) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 21.756711ms) + Feb 25 09:47:32.037: INFO: (16) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 23.471012ms) + Feb 25 09:47:32.037: INFO: (16) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 23.209219ms) + Feb 25 09:47:32.038: INFO: (16) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 23.625235ms) + Feb 25 09:47:32.038: INFO: (16) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 24.498494ms) + Feb 25 09:47:32.040: INFO: (16) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname1/proxy/: foo (200; 25.965387ms) + Feb 25 09:47:32.062: INFO: (17) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 21.538867ms) + Feb 25 09:47:32.064: INFO: (17) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 21.903374ms) + Feb 25 09:47:32.064: INFO: (17) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 23.440285ms) + Feb 25 09:47:32.067: INFO: (17) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 24.517662ms) + Feb 25 09:47:32.067: INFO: (17) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname1/proxy/: tls baz (200; 25.886767ms) + Feb 25 09:47:32.067: INFO: (17) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test (200; 26.362812ms) + Feb 25 09:47:32.068: INFO: (17) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 26.605805ms) + Feb 25 09:47:32.068: INFO: (17) /api/v1/namespaces/proxy-7283/services/proxy-service-b9nvz:portname2/proxy/: bar (200; 27.444871ms) + Feb 25 09:47:32.068: INFO: (17) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 26.68948ms) + Feb 25 09:47:32.069: INFO: (17) /api/v1/namespaces/proxy-7283/services/https:proxy-service-b9nvz:tlsportname2/proxy/: tls qux (200; 28.713331ms) + Feb 25 09:47:32.070: INFO: (17) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 27.519819ms) + Feb 25 09:47:32.070: INFO: (17) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 28.952531ms) + Feb 25 09:47:32.070: INFO: (17) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 29.084565ms) + Feb 25 09:47:32.082: INFO: (18) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:460/proxy/: tls baz (200; 10.746554ms) + Feb 25 09:47:32.083: INFO: (18) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 12.234611ms) + Feb 25 09:47:32.084: INFO: (18) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 13.399517ms) + Feb 25 09:47:32.085: INFO: (18) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:462/proxy/: tls qux (200; 13.54368ms) + Feb 25 09:47:32.086: INFO: (18) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 14.701604ms) + Feb 25 09:47:32.087: INFO: (18) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 16.017447ms) + Feb 25 09:47:32.090: INFO: (18) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:162/proxy/: bar (200; 18.606878ms) + Feb 25 09:47:32.090: INFO: (18) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: test<... (200; 27.632653ms) + Feb 25 09:47:32.100: INFO: (18) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname2/proxy/: bar (200; 27.978582ms) + Feb 25 09:47:32.109: INFO: (19) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk/proxy/: test (200; 8.02944ms) + Feb 25 09:47:32.110: INFO: (19) /api/v1/namespaces/proxy-7283/pods/http:proxy-service-b9nvz-sfdqk:1080/proxy/: ... (200; 9.11425ms) + Feb 25 09:47:32.121: INFO: (19) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:160/proxy/: foo (200; 18.80755ms) + Feb 25 09:47:32.121: INFO: (19) /api/v1/namespaces/proxy-7283/services/http:proxy-service-b9nvz:portname1/proxy/: foo (200; 19.151867ms) + Feb 25 09:47:32.121: INFO: (19) /api/v1/namespaces/proxy-7283/pods/proxy-service-b9nvz-sfdqk:1080/proxy/: test<... (200; 18.797144ms) + Feb 25 09:47:32.122: INFO: (19) /api/v1/namespaces/proxy-7283/pods/https:proxy-service-b9nvz-sfdqk:443/proxy/: >> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename daemonsets 12/14/22 15:37:01.09 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:37:01.125 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:37:01.13 -[BeforeEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:145 -[It] should verify changes to a daemon set status [Conformance] - test/e2e/apps/daemon_set.go:861 -STEP: Creating simple DaemonSet "daemon-set" 12/14/22 15:37:01.176 -STEP: Check that daemon pods launch on every node of the cluster. 12/14/22 15:37:01.188 -Dec 14 15:37:01.207: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 15:37:01.207: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 15:37:02.221: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 15:37:02.221: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 15:37:03.219: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 15:37:03.219: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 15:37:04.219: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 15:37:04.219: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 15:37:05.219: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 15:37:05.219: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 15:37:06.221: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 15:37:06.221: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 15:37:07.218: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 15:37:07.218: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 15:37:08.223: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 -Dec 14 15:37:08.223: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 15:37:09.220: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 -Dec 14 15:37:09.220: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set -STEP: Getting /status 12/14/22 15:37:09.224 -Dec 14 15:37:09.234: INFO: Daemon Set daemon-set has Conditions: [] -STEP: updating the DaemonSet Status 12/14/22 15:37:09.234 -Dec 14 15:37:09.247: INFO: updatedStatus.Conditions: []v1.DaemonSetCondition{v1.DaemonSetCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} -STEP: watching for the daemon set status to be updated 12/14/22 15:37:09.248 -Dec 14 15:37:09.252: INFO: Observed &DaemonSet event: ADDED -Dec 14 15:37:09.252: INFO: Observed &DaemonSet event: MODIFIED -Dec 14 15:37:09.252: INFO: Observed &DaemonSet event: MODIFIED -Dec 14 15:37:09.252: INFO: Observed &DaemonSet event: MODIFIED -Dec 14 15:37:09.253: INFO: Observed &DaemonSet event: MODIFIED -Dec 14 15:37:09.253: INFO: Observed &DaemonSet event: MODIFIED -Dec 14 15:37:09.253: INFO: Found daemon set daemon-set in namespace daemonsets-4927 with labels: map[daemonset-name:daemon-set] annotations: map[deprecated.daemonset.template.generation:1] & Conditions: [{StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] -Dec 14 15:37:09.253: INFO: Daemon set daemon-set has an updated status -STEP: patching the DaemonSet Status 12/14/22 15:37:09.253 -STEP: watching for the daemon set status to be patched 12/14/22 15:37:09.267 -Dec 14 15:37:09.270: INFO: Observed &DaemonSet event: ADDED -Dec 14 15:37:09.270: INFO: Observed &DaemonSet event: MODIFIED -Dec 14 15:37:09.270: INFO: Observed &DaemonSet event: MODIFIED -Dec 14 15:37:09.270: INFO: Observed &DaemonSet event: MODIFIED -Dec 14 15:37:09.271: INFO: Observed &DaemonSet event: MODIFIED -Dec 14 15:37:09.271: INFO: Observed &DaemonSet event: MODIFIED -Dec 14 15:37:09.271: INFO: Observed daemon set daemon-set in namespace daemonsets-4927 with annotations: map[deprecated.daemonset.template.generation:1] & Conditions: [{StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] -Dec 14 15:37:09.272: INFO: Observed &DaemonSet event: MODIFIED -Dec 14 15:37:09.272: INFO: Found daemon set daemon-set in namespace daemonsets-4927 with labels: map[daemonset-name:daemon-set] annotations: map[deprecated.daemonset.template.generation:1] & Conditions: [{StatusPatched True 0001-01-01 00:00:00 +0000 UTC }] -Dec 14 15:37:09.272: INFO: Daemon set daemon-set has a patched status -[AfterEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:110 -STEP: Deleting DaemonSet "daemon-set" 12/14/22 15:37:09.287 -STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-4927, will wait for the garbage collector to delete the pods 12/14/22 15:37:09.287 -Dec 14 15:37:09.354: INFO: Deleting DaemonSet.extensions daemon-set took: 10.809626ms -Dec 14 15:37:09.456: INFO: Terminating DaemonSet.extensions daemon-set pods took: 101.473331ms -Dec 14 15:37:12.266: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 15:37:12.266: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set -Dec 14 15:37:12.272: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"2281"},"items":null} - -Dec 14 15:37:12.277: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"2281"},"items":null} - -[AfterEach] [sig-apps] Daemon set [Serial] +STEP: Creating a kubernetes client 02/25/23 09:47:34.541 +Feb 25 09:47:34.541: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 09:47:34.544 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:34.57 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:34.577 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 +[It] should add annotations for pods in rc [Conformance] + test/e2e/kubectl/kubectl.go:1650 +STEP: creating Agnhost RC 02/25/23 09:47:34.582 +Feb 25 09:47:34.583: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1457 create -f -' +Feb 25 09:47:36.582: INFO: stderr: "" +Feb 25 09:47:36.582: INFO: stdout: "replicationcontroller/agnhost-primary created\n" +STEP: Waiting for Agnhost primary to start. 02/25/23 09:47:36.582 +Feb 25 09:47:37.591: INFO: Selector matched 1 pods for map[app:agnhost] +Feb 25 09:47:37.591: INFO: Found 0 / 1 +Feb 25 09:47:38.590: INFO: Selector matched 1 pods for map[app:agnhost] +Feb 25 09:47:38.590: INFO: Found 1 / 1 +Feb 25 09:47:38.590: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 +STEP: patching all pods 02/25/23 09:47:38.59 +Feb 25 09:47:38.595: INFO: Selector matched 1 pods for map[app:agnhost] +Feb 25 09:47:38.595: INFO: ForEach: Found 1 pods from the filter. Now looping through them. +Feb 25 09:47:38.596: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1457 patch pod agnhost-primary-rp7hk -p {"metadata":{"annotations":{"x":"y"}}}' +Feb 25 09:47:38.753: INFO: stderr: "" +Feb 25 09:47:38.753: INFO: stdout: "pod/agnhost-primary-rp7hk patched\n" +STEP: checking annotations 02/25/23 09:47:38.753 +Feb 25 09:47:38.760: INFO: Selector matched 1 pods for map[app:agnhost] +Feb 25 09:47:38.760: INFO: ForEach: Found 1 pods from the filter. Now looping through them. +[AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 -Dec 14 15:37:12.301: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "daemonsets-4927" for this suite. 12/14/22 15:37:12.307 -{"msg":"PASSED [sig-apps] Daemon set [Serial] should verify changes to a daemon set status [Conformance]","completed":5,"skipped":23,"failed":0} +Feb 25 09:47:38.761: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-1457" for this suite. 02/25/23 09:47:38.768 +{"msg":"PASSED [sig-cli] Kubectl client Kubectl patch should add annotations for pods in rc [Conformance]","completed":5,"skipped":191,"failed":0} ------------------------------ -• [SLOW TEST] [11.232 seconds] -[sig-apps] Daemon set [Serial] -test/e2e/apps/framework.go:23 - should verify changes to a daemon set status [Conformance] - test/e2e/apps/daemon_set.go:861 +• [4.237 seconds] +[sig-cli] Kubectl client +test/e2e/kubectl/framework.go:23 + Kubectl patch + test/e2e/kubectl/kubectl.go:1644 + should add annotations for pods in rc [Conformance] + test/e2e/kubectl/kubectl.go:1650 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Daemon set [Serial] + [BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:37:01.088 - Dec 14 15:37:01.088: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename daemonsets 12/14/22 15:37:01.09 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:37:01.125 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:37:01.13 - [BeforeEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:145 - [It] should verify changes to a daemon set status [Conformance] - test/e2e/apps/daemon_set.go:861 - STEP: Creating simple DaemonSet "daemon-set" 12/14/22 15:37:01.176 - STEP: Check that daemon pods launch on every node of the cluster. 12/14/22 15:37:01.188 - Dec 14 15:37:01.207: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 15:37:01.207: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 15:37:02.221: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 15:37:02.221: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 15:37:03.219: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 15:37:03.219: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 15:37:04.219: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 15:37:04.219: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 15:37:05.219: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 15:37:05.219: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 15:37:06.221: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 15:37:06.221: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 15:37:07.218: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 15:37:07.218: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 15:37:08.223: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 - Dec 14 15:37:08.223: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 15:37:09.220: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 - Dec 14 15:37:09.220: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set - STEP: Getting /status 12/14/22 15:37:09.224 - Dec 14 15:37:09.234: INFO: Daemon Set daemon-set has Conditions: [] - STEP: updating the DaemonSet Status 12/14/22 15:37:09.234 - Dec 14 15:37:09.247: INFO: updatedStatus.Conditions: []v1.DaemonSetCondition{v1.DaemonSetCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} - STEP: watching for the daemon set status to be updated 12/14/22 15:37:09.248 - Dec 14 15:37:09.252: INFO: Observed &DaemonSet event: ADDED - Dec 14 15:37:09.252: INFO: Observed &DaemonSet event: MODIFIED - Dec 14 15:37:09.252: INFO: Observed &DaemonSet event: MODIFIED - Dec 14 15:37:09.252: INFO: Observed &DaemonSet event: MODIFIED - Dec 14 15:37:09.253: INFO: Observed &DaemonSet event: MODIFIED - Dec 14 15:37:09.253: INFO: Observed &DaemonSet event: MODIFIED - Dec 14 15:37:09.253: INFO: Found daemon set daemon-set in namespace daemonsets-4927 with labels: map[daemonset-name:daemon-set] annotations: map[deprecated.daemonset.template.generation:1] & Conditions: [{StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] - Dec 14 15:37:09.253: INFO: Daemon set daemon-set has an updated status - STEP: patching the DaemonSet Status 12/14/22 15:37:09.253 - STEP: watching for the daemon set status to be patched 12/14/22 15:37:09.267 - Dec 14 15:37:09.270: INFO: Observed &DaemonSet event: ADDED - Dec 14 15:37:09.270: INFO: Observed &DaemonSet event: MODIFIED - Dec 14 15:37:09.270: INFO: Observed &DaemonSet event: MODIFIED - Dec 14 15:37:09.270: INFO: Observed &DaemonSet event: MODIFIED - Dec 14 15:37:09.271: INFO: Observed &DaemonSet event: MODIFIED - Dec 14 15:37:09.271: INFO: Observed &DaemonSet event: MODIFIED - Dec 14 15:37:09.271: INFO: Observed daemon set daemon-set in namespace daemonsets-4927 with annotations: map[deprecated.daemonset.template.generation:1] & Conditions: [{StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] - Dec 14 15:37:09.272: INFO: Observed &DaemonSet event: MODIFIED - Dec 14 15:37:09.272: INFO: Found daemon set daemon-set in namespace daemonsets-4927 with labels: map[daemonset-name:daemon-set] annotations: map[deprecated.daemonset.template.generation:1] & Conditions: [{StatusPatched True 0001-01-01 00:00:00 +0000 UTC }] - Dec 14 15:37:09.272: INFO: Daemon set daemon-set has a patched status - [AfterEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:110 - STEP: Deleting DaemonSet "daemon-set" 12/14/22 15:37:09.287 - STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-4927, will wait for the garbage collector to delete the pods 12/14/22 15:37:09.287 - Dec 14 15:37:09.354: INFO: Deleting DaemonSet.extensions daemon-set took: 10.809626ms - Dec 14 15:37:09.456: INFO: Terminating DaemonSet.extensions daemon-set pods took: 101.473331ms - Dec 14 15:37:12.266: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 15:37:12.266: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set - Dec 14 15:37:12.272: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"2281"},"items":null} - - Dec 14 15:37:12.277: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"2281"},"items":null} - - [AfterEach] [sig-apps] Daemon set [Serial] + STEP: Creating a kubernetes client 02/25/23 09:47:34.541 + Feb 25 09:47:34.541: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 09:47:34.544 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:34.57 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:34.577 + [BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 + [It] should add annotations for pods in rc [Conformance] + test/e2e/kubectl/kubectl.go:1650 + STEP: creating Agnhost RC 02/25/23 09:47:34.582 + Feb 25 09:47:34.583: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1457 create -f -' + Feb 25 09:47:36.582: INFO: stderr: "" + Feb 25 09:47:36.582: INFO: stdout: "replicationcontroller/agnhost-primary created\n" + STEP: Waiting for Agnhost primary to start. 02/25/23 09:47:36.582 + Feb 25 09:47:37.591: INFO: Selector matched 1 pods for map[app:agnhost] + Feb 25 09:47:37.591: INFO: Found 0 / 1 + Feb 25 09:47:38.590: INFO: Selector matched 1 pods for map[app:agnhost] + Feb 25 09:47:38.590: INFO: Found 1 / 1 + Feb 25 09:47:38.590: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 + STEP: patching all pods 02/25/23 09:47:38.59 + Feb 25 09:47:38.595: INFO: Selector matched 1 pods for map[app:agnhost] + Feb 25 09:47:38.595: INFO: ForEach: Found 1 pods from the filter. Now looping through them. + Feb 25 09:47:38.596: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1457 patch pod agnhost-primary-rp7hk -p {"metadata":{"annotations":{"x":"y"}}}' + Feb 25 09:47:38.753: INFO: stderr: "" + Feb 25 09:47:38.753: INFO: stdout: "pod/agnhost-primary-rp7hk patched\n" + STEP: checking annotations 02/25/23 09:47:38.753 + Feb 25 09:47:38.760: INFO: Selector matched 1 pods for map[app:agnhost] + Feb 25 09:47:38.760: INFO: ForEach: Found 1 pods from the filter. Now looping through them. + [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 15:37:12.301: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "daemonsets-4927" for this suite. 12/14/22 15:37:12.307 + Feb 25 09:47:38.761: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-1457" for this suite. 02/25/23 09:47:38.768 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSS +SSSSSSSSSS ------------------------------ -[sig-node] Variable Expansion - should fail substituting values in a volume subpath with absolute path [Slow] [Conformance] - test/e2e/common/node/expansion.go:185 -[BeforeEach] [sig-node] Variable Expansion +[sig-storage] EmptyDir wrapper volumes + should not conflict [Conformance] + test/e2e/storage/empty_dir_wrapper.go:67 +[BeforeEach] [sig-storage] EmptyDir wrapper volumes test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:37:12.322 -Dec 14 15:37:12.322: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename var-expansion 12/14/22 15:37:12.327 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:37:12.357 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:37:12.36 -[It] should fail substituting values in a volume subpath with absolute path [Slow] [Conformance] - test/e2e/common/node/expansion.go:185 -Dec 14 15:37:12.379: INFO: Waiting up to 2m0s for pod "var-expansion-d7208d96-4b5b-4921-be62-ee781af25d51" in namespace "var-expansion-4159" to be "container 0 failed with reason CreateContainerConfigError" -Dec 14 15:37:12.387: INFO: Pod "var-expansion-d7208d96-4b5b-4921-be62-ee781af25d51": Phase="Pending", Reason="", readiness=false. Elapsed: 7.369396ms -Dec 14 15:37:14.397: INFO: Pod "var-expansion-d7208d96-4b5b-4921-be62-ee781af25d51": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017393679s -Dec 14 15:37:16.399: INFO: Pod "var-expansion-d7208d96-4b5b-4921-be62-ee781af25d51": Phase="Pending", Reason="", readiness=false. Elapsed: 4.019859036s -Dec 14 15:37:16.399: INFO: Pod "var-expansion-d7208d96-4b5b-4921-be62-ee781af25d51" satisfied condition "container 0 failed with reason CreateContainerConfigError" -Dec 14 15:37:16.399: INFO: Deleting pod "var-expansion-d7208d96-4b5b-4921-be62-ee781af25d51" in namespace "var-expansion-4159" -Dec 14 15:37:16.411: INFO: Wait up to 5m0s for pod "var-expansion-d7208d96-4b5b-4921-be62-ee781af25d51" to be fully deleted -[AfterEach] [sig-node] Variable Expansion +STEP: Creating a kubernetes client 02/25/23 09:47:38.779 +Feb 25 09:47:38.780: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir-wrapper 02/25/23 09:47:38.784 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:38.833 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:38.839 +[It] should not conflict [Conformance] + test/e2e/storage/empty_dir_wrapper.go:67 +Feb 25 09:47:38.887: INFO: Waiting up to 5m0s for pod "pod-secrets-28c18aeb-dead-4a66-9e49-8e89a63d2ab4" in namespace "emptydir-wrapper-3310" to be "running and ready" +Feb 25 09:47:38.893: INFO: Pod "pod-secrets-28c18aeb-dead-4a66-9e49-8e89a63d2ab4": Phase="Pending", Reason="", readiness=false. Elapsed: 5.918666ms +Feb 25 09:47:38.893: INFO: The phase of Pod pod-secrets-28c18aeb-dead-4a66-9e49-8e89a63d2ab4 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 09:47:40.902: INFO: Pod "pod-secrets-28c18aeb-dead-4a66-9e49-8e89a63d2ab4": Phase="Running", Reason="", readiness=true. Elapsed: 2.014824841s +Feb 25 09:47:40.902: INFO: The phase of Pod pod-secrets-28c18aeb-dead-4a66-9e49-8e89a63d2ab4 is Running (Ready = true) +Feb 25 09:47:40.902: INFO: Pod "pod-secrets-28c18aeb-dead-4a66-9e49-8e89a63d2ab4" satisfied condition "running and ready" +STEP: Cleaning up the secret 02/25/23 09:47:40.908 +STEP: Cleaning up the configmap 02/25/23 09:47:40.919 +STEP: Cleaning up the pod 02/25/23 09:47:40.929 +[AfterEach] [sig-storage] EmptyDir wrapper volumes test/e2e/framework/framework.go:187 -Dec 14 15:37:18.425: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "var-expansion-4159" for this suite. 12/14/22 15:37:18.432 -{"msg":"PASSED [sig-node] Variable Expansion should fail substituting values in a volume subpath with absolute path [Slow] [Conformance]","completed":6,"skipped":32,"failed":0} +Feb 25 09:47:40.967: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-wrapper-3310" for this suite. 02/25/23 09:47:40.977 +{"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not conflict [Conformance]","completed":6,"skipped":201,"failed":0} ------------------------------ -• [SLOW TEST] [6.119 seconds] -[sig-node] Variable Expansion -test/e2e/common/node/framework.go:23 - should fail substituting values in a volume subpath with absolute path [Slow] [Conformance] - test/e2e/common/node/expansion.go:185 +• [2.209 seconds] +[sig-storage] EmptyDir wrapper volumes +test/e2e/storage/utils/framework.go:23 + should not conflict [Conformance] + test/e2e/storage/empty_dir_wrapper.go:67 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Variable Expansion + [BeforeEach] [sig-storage] EmptyDir wrapper volumes test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:37:12.322 - Dec 14 15:37:12.322: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename var-expansion 12/14/22 15:37:12.327 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:37:12.357 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:37:12.36 - [It] should fail substituting values in a volume subpath with absolute path [Slow] [Conformance] - test/e2e/common/node/expansion.go:185 - Dec 14 15:37:12.379: INFO: Waiting up to 2m0s for pod "var-expansion-d7208d96-4b5b-4921-be62-ee781af25d51" in namespace "var-expansion-4159" to be "container 0 failed with reason CreateContainerConfigError" - Dec 14 15:37:12.387: INFO: Pod "var-expansion-d7208d96-4b5b-4921-be62-ee781af25d51": Phase="Pending", Reason="", readiness=false. Elapsed: 7.369396ms - Dec 14 15:37:14.397: INFO: Pod "var-expansion-d7208d96-4b5b-4921-be62-ee781af25d51": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017393679s - Dec 14 15:37:16.399: INFO: Pod "var-expansion-d7208d96-4b5b-4921-be62-ee781af25d51": Phase="Pending", Reason="", readiness=false. Elapsed: 4.019859036s - Dec 14 15:37:16.399: INFO: Pod "var-expansion-d7208d96-4b5b-4921-be62-ee781af25d51" satisfied condition "container 0 failed with reason CreateContainerConfigError" - Dec 14 15:37:16.399: INFO: Deleting pod "var-expansion-d7208d96-4b5b-4921-be62-ee781af25d51" in namespace "var-expansion-4159" - Dec 14 15:37:16.411: INFO: Wait up to 5m0s for pod "var-expansion-d7208d96-4b5b-4921-be62-ee781af25d51" to be fully deleted - [AfterEach] [sig-node] Variable Expansion + STEP: Creating a kubernetes client 02/25/23 09:47:38.779 + Feb 25 09:47:38.780: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir-wrapper 02/25/23 09:47:38.784 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:38.833 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:38.839 + [It] should not conflict [Conformance] + test/e2e/storage/empty_dir_wrapper.go:67 + Feb 25 09:47:38.887: INFO: Waiting up to 5m0s for pod "pod-secrets-28c18aeb-dead-4a66-9e49-8e89a63d2ab4" in namespace "emptydir-wrapper-3310" to be "running and ready" + Feb 25 09:47:38.893: INFO: Pod "pod-secrets-28c18aeb-dead-4a66-9e49-8e89a63d2ab4": Phase="Pending", Reason="", readiness=false. Elapsed: 5.918666ms + Feb 25 09:47:38.893: INFO: The phase of Pod pod-secrets-28c18aeb-dead-4a66-9e49-8e89a63d2ab4 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 09:47:40.902: INFO: Pod "pod-secrets-28c18aeb-dead-4a66-9e49-8e89a63d2ab4": Phase="Running", Reason="", readiness=true. Elapsed: 2.014824841s + Feb 25 09:47:40.902: INFO: The phase of Pod pod-secrets-28c18aeb-dead-4a66-9e49-8e89a63d2ab4 is Running (Ready = true) + Feb 25 09:47:40.902: INFO: Pod "pod-secrets-28c18aeb-dead-4a66-9e49-8e89a63d2ab4" satisfied condition "running and ready" + STEP: Cleaning up the secret 02/25/23 09:47:40.908 + STEP: Cleaning up the configmap 02/25/23 09:47:40.919 + STEP: Cleaning up the pod 02/25/23 09:47:40.929 + [AfterEach] [sig-storage] EmptyDir wrapper volumes test/e2e/framework/framework.go:187 - Dec 14 15:37:18.425: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "var-expansion-4159" for this suite. 12/14/22 15:37:18.432 + Feb 25 09:47:40.967: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-wrapper-3310" for this suite. 02/25/23 09:47:40.977 << End Captured GinkgoWriter Output ------------------------------ -S +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-scheduling] SchedulerPredicates [Serial] - validates that NodeSelector is respected if not matching [Conformance] - test/e2e/scheduling/predicates.go:438 -[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] +[sig-apps] Deployment + should validate Deployment Status endpoints [Conformance] + test/e2e/apps/deployment.go:479 +[BeforeEach] [sig-apps] Deployment test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:37:18.445 -Dec 14 15:37:18.445: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename sched-pred 12/14/22 15:37:18.447 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:37:18.47 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:37:18.475 -[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:92 -Dec 14 15:37:18.478: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready -Dec 14 15:37:18.490: INFO: Waiting for terminating namespaces to be deleted... -Dec 14 15:37:18.494: INFO: -Logging pods the apiserver thinks is on node kiedai4iakae-1 before test -Dec 14 15:37:18.504: INFO: kube-flannel-ds-d6vqv from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.504: INFO: Container kube-flannel ready: true, restart count 0 -Dec 14 15:37:18.504: INFO: kube-addon-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.504: INFO: Container kube-addon-manager ready: true, restart count 1 -Dec 14 15:37:18.504: INFO: kube-apiserver-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.504: INFO: Container kube-apiserver ready: true, restart count 1 -Dec 14 15:37:18.504: INFO: kube-controller-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:29:20 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.504: INFO: Container kube-controller-manager ready: true, restart count 2 -Dec 14 15:37:18.504: INFO: kube-proxy-fs5fm from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.504: INFO: Container kube-proxy ready: true, restart count 0 -Dec 14 15:37:18.504: INFO: kube-scheduler-kiedai4iakae-1 from kube-system started at 2022-12-14 15:25:59 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.504: INFO: Container kube-scheduler ready: true, restart count 2 -Dec 14 15:37:18.504: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-z7zks from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 15:37:18.504: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 15:37:18.504: INFO: Container systemd-logs ready: true, restart count 0 -Dec 14 15:37:18.504: INFO: -Logging pods the apiserver thinks is on node kiedai4iakae-2 before test -Dec 14 15:37:18.512: INFO: kube-flannel-ds-nvdqm from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.512: INFO: Container kube-flannel ready: true, restart count 0 -Dec 14 15:37:18.512: INFO: coredns-565d847f94-xhv9r from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.512: INFO: Container coredns ready: true, restart count 0 -Dec 14 15:37:18.512: INFO: kube-addon-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.512: INFO: Container kube-addon-manager ready: true, restart count 1 -Dec 14 15:37:18.512: INFO: kube-apiserver-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.512: INFO: Container kube-apiserver ready: true, restart count 1 -Dec 14 15:37:18.512: INFO: kube-controller-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.512: INFO: Container kube-controller-manager ready: true, restart count 1 -Dec 14 15:37:18.512: INFO: kube-proxy-dqjw5 from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.512: INFO: Container kube-proxy ready: true, restart count 0 -Dec 14 15:37:18.512: INFO: kube-scheduler-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.512: INFO: Container kube-scheduler ready: true, restart count 2 -Dec 14 15:37:18.512: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-cgz78 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 15:37:18.512: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 15:37:18.512: INFO: Container systemd-logs ready: true, restart count 0 -Dec 14 15:37:18.512: INFO: -Logging pods the apiserver thinks is on node kiedai4iakae-3 before test -Dec 14 15:37:18.522: INFO: kube-flannel-ds-j6hx4 from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.522: INFO: Container kube-flannel ready: true, restart count 0 -Dec 14 15:37:18.522: INFO: coredns-565d847f94-6tl8c from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.522: INFO: Container coredns ready: true, restart count 0 -Dec 14 15:37:18.522: INFO: kube-proxy-8kpmb from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.522: INFO: Container kube-proxy ready: true, restart count 0 -Dec 14 15:37:18.522: INFO: sonobuoy from sonobuoy started at 2022-12-14 15:35:59 +0000 UTC (1 container statuses recorded) -Dec 14 15:37:18.522: INFO: Container kube-sonobuoy ready: true, restart count 0 -Dec 14 15:37:18.522: INFO: sonobuoy-e2e-job-177ae5e9cdd34bac from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 15:37:18.522: INFO: Container e2e ready: true, restart count 0 -Dec 14 15:37:18.522: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 15:37:18.522: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-2fkm5 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 15:37:18.523: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 15:37:18.523: INFO: Container systemd-logs ready: true, restart count 0 -[It] validates that NodeSelector is respected if not matching [Conformance] - test/e2e/scheduling/predicates.go:438 -STEP: Trying to schedule Pod with nonempty NodeSelector. 12/14/22 15:37:18.523 -STEP: Considering event: -Type = [Warning], Name = [restricted-pod.1730b2f7aaf439ea], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match Pod's node affinity/selector. preemption: 0/3 nodes are available: 3 Preemption is not helpful for scheduling.] 12/14/22 15:37:18.567 -[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] +STEP: Creating a kubernetes client 02/25/23 09:47:41.003 +Feb 25 09:47:41.003: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename deployment 02/25/23 09:47:41.005 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:41.056 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:41.06 +[BeforeEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:91 +[It] should validate Deployment Status endpoints [Conformance] + test/e2e/apps/deployment.go:479 +STEP: creating a Deployment 02/25/23 09:47:41.068 +Feb 25 09:47:41.068: INFO: Creating simple deployment test-deployment-hkpfs +Feb 25 09:47:41.085: INFO: new replicaset for deployment "test-deployment-hkpfs" is yet to be created +Feb 25 09:47:43.110: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-deployment-hkpfs-777898ffcc\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 09:47:45.124: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-deployment-hkpfs-777898ffcc\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 09:47:47.121: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-deployment-hkpfs-777898ffcc\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 09:47:49.127: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-deployment-hkpfs-777898ffcc\" is progressing."}}, CollisionCount:(*int32)(nil)} +STEP: Getting /status 02/25/23 09:47:51.128 +Feb 25 09:47:51.140: INFO: Deployment test-deployment-hkpfs has Conditions: [{Available True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:49 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-hkpfs-777898ffcc" has successfully progressed.}] +STEP: updating Deployment Status 02/25/23 09:47:51.14 +Feb 25 09:47:51.158: INFO: updatedStatus.Conditions: []v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 49, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 49, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 49, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"NewReplicaSetAvailable", Message:"ReplicaSet \"test-deployment-hkpfs-777898ffcc\" has successfully progressed."}, v1.DeploymentCondition{Type:"StatusUpdate", Status:"True", LastUpdateTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} +STEP: watching for the Deployment status to be updated 02/25/23 09:47:51.158 +Feb 25 09:47:51.162: INFO: Observed &Deployment event: ADDED +Feb 25 09:47:51.163: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-hkpfs-777898ffcc"} +Feb 25 09:47:51.163: INFO: Observed &Deployment event: MODIFIED +Feb 25 09:47:51.163: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-hkpfs-777898ffcc"} +Feb 25 09:47:51.164: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} +Feb 25 09:47:51.164: INFO: Observed &Deployment event: MODIFIED +Feb 25 09:47:51.164: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} +Feb 25 09:47:51.164: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC ReplicaSetUpdated ReplicaSet "test-deployment-hkpfs-777898ffcc" is progressing.} +Feb 25 09:47:51.164: INFO: Observed &Deployment event: MODIFIED +Feb 25 09:47:51.165: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:49 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} +Feb 25 09:47:51.165: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-hkpfs-777898ffcc" has successfully progressed.} +Feb 25 09:47:51.165: INFO: Observed &Deployment event: MODIFIED +Feb 25 09:47:51.165: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:49 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} +Feb 25 09:47:51.165: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-hkpfs-777898ffcc" has successfully progressed.} +Feb 25 09:47:51.165: INFO: Found Deployment test-deployment-hkpfs in namespace deployment-2532 with labels: map[e2e:testing name:httpd] annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} +Feb 25 09:47:51.165: INFO: Deployment test-deployment-hkpfs has an updated status +STEP: patching the Statefulset Status 02/25/23 09:47:51.165 +Feb 25 09:47:51.166: INFO: Patch payload: {"status":{"conditions":[{"type":"StatusPatched","status":"True"}]}} +Feb 25 09:47:51.175: INFO: Patched status conditions: []v1.DeploymentCondition{v1.DeploymentCondition{Type:"StatusPatched", Status:"True", LastUpdateTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"", Message:""}} +STEP: watching for the Deployment status to be patched 02/25/23 09:47:51.175 +Feb 25 09:47:51.178: INFO: Observed &Deployment event: ADDED +Feb 25 09:47:51.178: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-hkpfs-777898ffcc"} +Feb 25 09:47:51.178: INFO: Observed &Deployment event: MODIFIED +Feb 25 09:47:51.178: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-hkpfs-777898ffcc"} +Feb 25 09:47:51.178: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} +Feb 25 09:47:51.179: INFO: Observed &Deployment event: MODIFIED +Feb 25 09:47:51.179: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} +Feb 25 09:47:51.179: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC ReplicaSetUpdated ReplicaSet "test-deployment-hkpfs-777898ffcc" is progressing.} +Feb 25 09:47:51.179: INFO: Observed &Deployment event: MODIFIED +Feb 25 09:47:51.180: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:49 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} +Feb 25 09:47:51.180: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-hkpfs-777898ffcc" has successfully progressed.} +Feb 25 09:47:51.180: INFO: Observed &Deployment event: MODIFIED +Feb 25 09:47:51.180: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:49 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} +Feb 25 09:47:51.180: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-hkpfs-777898ffcc" has successfully progressed.} +Feb 25 09:47:51.180: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} +Feb 25 09:47:51.180: INFO: Observed &Deployment event: MODIFIED +Feb 25 09:47:51.180: INFO: Found deployment test-deployment-hkpfs in namespace deployment-2532 with labels: map[e2e:testing name:httpd] annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {StatusPatched True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } +Feb 25 09:47:51.180: INFO: Deployment test-deployment-hkpfs has a patched status +[AfterEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:84 +Feb 25 09:47:51.193: INFO: Deployment "test-deployment-hkpfs": +&Deployment{ObjectMeta:{test-deployment-hkpfs deployment-2532 bc5f99c9-0e74-4e89-9947-ffd92c4558f5 8913 1 2023-02-25 09:47:41 +0000 UTC map[e2e:testing name:httpd] map[deployment.kubernetes.io/revision:1] [] [] [{e2e.test Update apps/v1 2023-02-25 09:47:41 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:e2e":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:e2e":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {e2e.test Update apps/v1 2023-02-25 09:47:51 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"StatusPatched\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:status":{},"f:type":{}}}}} status} {kube-controller-manager Update apps/v1 2023-02-25 09:47:51 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{e2e: testing,name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[e2e:testing name:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc001568708 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:StatusPatched,Status:True,Reason:,Message:,LastUpdateTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:0001-01-01 00:00:00 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:FoundNewReplicaSet,Message:Found new replica set "test-deployment-hkpfs-777898ffcc",LastUpdateTime:2023-02-25 09:47:51 +0000 UTC,LastTransitionTime:2023-02-25 09:47:51 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} + +Feb 25 09:47:51.201: INFO: New ReplicaSet "test-deployment-hkpfs-777898ffcc" of Deployment "test-deployment-hkpfs": +&ReplicaSet{ObjectMeta:{test-deployment-hkpfs-777898ffcc deployment-2532 dcf8ee03-84e2-4c62-8743-a4772da4d093 8905 1 2023-02-25 09:47:41 +0000 UTC map[e2e:testing name:httpd pod-template-hash:777898ffcc] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-deployment-hkpfs bc5f99c9-0e74-4e89-9947-ffd92c4558f5 0xc001589237 0xc001589238}] [] [{kube-controller-manager Update apps/v1 2023-02-25 09:47:41 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:e2e":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"bc5f99c9-0e74-4e89-9947-ffd92c4558f5\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:e2e":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:47:49 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{e2e: testing,name: httpd,pod-template-hash: 777898ffcc,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[e2e:testing name:httpd pod-template-hash:777898ffcc] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0015892e8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} +Feb 25 09:47:51.210: INFO: Pod "test-deployment-hkpfs-777898ffcc-djm5x" is available: +&Pod{ObjectMeta:{test-deployment-hkpfs-777898ffcc-djm5x test-deployment-hkpfs-777898ffcc- deployment-2532 c3dbe72b-33e2-4212-8944-f9386f7a5e75 8904 0 2023-02-25 09:47:41 +0000 UTC map[e2e:testing name:httpd pod-template-hash:777898ffcc] map[] [{apps/v1 ReplicaSet test-deployment-hkpfs-777898ffcc dcf8ee03-84e2-4c62-8743-a4772da4d093 0xc001589687 0xc001589688}] [] [{kube-controller-manager Update v1 2023-02-25 09:47:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:e2e":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"dcf8ee03-84e2-4c62-8743-a4772da4d093\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:47:49 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.194\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-rdw9z,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rdw9z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:47:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:47:49 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:47:49 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:47:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:10.233.66.194,StartTime:2023-02-25 09:47:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:47:49 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://b2c62d4cbc88f0f4d735b9954601c7c9a5492b3659bc27ac13bb0657cf797194,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.194,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +[AfterEach] [sig-apps] Deployment test/e2e/framework/framework.go:187 -Dec 14 15:37:19.562: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "sched-pred-6452" for this suite. 12/14/22 15:37:19.57 -[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:83 -{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching [Conformance]","completed":7,"skipped":33,"failed":0} +Feb 25 09:47:51.210: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "deployment-2532" for this suite. 02/25/23 09:47:51.216 +{"msg":"PASSED [sig-apps] Deployment should validate Deployment Status endpoints [Conformance]","completed":7,"skipped":238,"failed":0} ------------------------------ -• [1.137 seconds] -[sig-scheduling] SchedulerPredicates [Serial] -test/e2e/scheduling/framework.go:40 - validates that NodeSelector is respected if not matching [Conformance] - test/e2e/scheduling/predicates.go:438 +• [SLOW TEST] [10.224 seconds] +[sig-apps] Deployment +test/e2e/apps/framework.go:23 + should validate Deployment Status endpoints [Conformance] + test/e2e/apps/deployment.go:479 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + [BeforeEach] [sig-apps] Deployment test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:37:18.445 - Dec 14 15:37:18.445: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename sched-pred 12/14/22 15:37:18.447 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:37:18.47 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:37:18.475 - [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:92 - Dec 14 15:37:18.478: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready - Dec 14 15:37:18.490: INFO: Waiting for terminating namespaces to be deleted... - Dec 14 15:37:18.494: INFO: - Logging pods the apiserver thinks is on node kiedai4iakae-1 before test - Dec 14 15:37:18.504: INFO: kube-flannel-ds-d6vqv from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.504: INFO: Container kube-flannel ready: true, restart count 0 - Dec 14 15:37:18.504: INFO: kube-addon-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.504: INFO: Container kube-addon-manager ready: true, restart count 1 - Dec 14 15:37:18.504: INFO: kube-apiserver-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.504: INFO: Container kube-apiserver ready: true, restart count 1 - Dec 14 15:37:18.504: INFO: kube-controller-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:29:20 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.504: INFO: Container kube-controller-manager ready: true, restart count 2 - Dec 14 15:37:18.504: INFO: kube-proxy-fs5fm from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.504: INFO: Container kube-proxy ready: true, restart count 0 - Dec 14 15:37:18.504: INFO: kube-scheduler-kiedai4iakae-1 from kube-system started at 2022-12-14 15:25:59 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.504: INFO: Container kube-scheduler ready: true, restart count 2 - Dec 14 15:37:18.504: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-z7zks from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 15:37:18.504: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 15:37:18.504: INFO: Container systemd-logs ready: true, restart count 0 - Dec 14 15:37:18.504: INFO: - Logging pods the apiserver thinks is on node kiedai4iakae-2 before test - Dec 14 15:37:18.512: INFO: kube-flannel-ds-nvdqm from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.512: INFO: Container kube-flannel ready: true, restart count 0 - Dec 14 15:37:18.512: INFO: coredns-565d847f94-xhv9r from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.512: INFO: Container coredns ready: true, restart count 0 - Dec 14 15:37:18.512: INFO: kube-addon-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.512: INFO: Container kube-addon-manager ready: true, restart count 1 - Dec 14 15:37:18.512: INFO: kube-apiserver-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.512: INFO: Container kube-apiserver ready: true, restart count 1 - Dec 14 15:37:18.512: INFO: kube-controller-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.512: INFO: Container kube-controller-manager ready: true, restart count 1 - Dec 14 15:37:18.512: INFO: kube-proxy-dqjw5 from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.512: INFO: Container kube-proxy ready: true, restart count 0 - Dec 14 15:37:18.512: INFO: kube-scheduler-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.512: INFO: Container kube-scheduler ready: true, restart count 2 - Dec 14 15:37:18.512: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-cgz78 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 15:37:18.512: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 15:37:18.512: INFO: Container systemd-logs ready: true, restart count 0 - Dec 14 15:37:18.512: INFO: - Logging pods the apiserver thinks is on node kiedai4iakae-3 before test - Dec 14 15:37:18.522: INFO: kube-flannel-ds-j6hx4 from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.522: INFO: Container kube-flannel ready: true, restart count 0 - Dec 14 15:37:18.522: INFO: coredns-565d847f94-6tl8c from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.522: INFO: Container coredns ready: true, restart count 0 - Dec 14 15:37:18.522: INFO: kube-proxy-8kpmb from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.522: INFO: Container kube-proxy ready: true, restart count 0 - Dec 14 15:37:18.522: INFO: sonobuoy from sonobuoy started at 2022-12-14 15:35:59 +0000 UTC (1 container statuses recorded) - Dec 14 15:37:18.522: INFO: Container kube-sonobuoy ready: true, restart count 0 - Dec 14 15:37:18.522: INFO: sonobuoy-e2e-job-177ae5e9cdd34bac from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 15:37:18.522: INFO: Container e2e ready: true, restart count 0 - Dec 14 15:37:18.522: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 15:37:18.522: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-2fkm5 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 15:37:18.523: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 15:37:18.523: INFO: Container systemd-logs ready: true, restart count 0 - [It] validates that NodeSelector is respected if not matching [Conformance] - test/e2e/scheduling/predicates.go:438 - STEP: Trying to schedule Pod with nonempty NodeSelector. 12/14/22 15:37:18.523 - STEP: Considering event: - Type = [Warning], Name = [restricted-pod.1730b2f7aaf439ea], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match Pod's node affinity/selector. preemption: 0/3 nodes are available: 3 Preemption is not helpful for scheduling.] 12/14/22 15:37:18.567 - [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + STEP: Creating a kubernetes client 02/25/23 09:47:41.003 + Feb 25 09:47:41.003: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename deployment 02/25/23 09:47:41.005 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:41.056 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:41.06 + [BeforeEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:91 + [It] should validate Deployment Status endpoints [Conformance] + test/e2e/apps/deployment.go:479 + STEP: creating a Deployment 02/25/23 09:47:41.068 + Feb 25 09:47:41.068: INFO: Creating simple deployment test-deployment-hkpfs + Feb 25 09:47:41.085: INFO: new replicaset for deployment "test-deployment-hkpfs" is yet to be created + Feb 25 09:47:43.110: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-deployment-hkpfs-777898ffcc\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 09:47:45.124: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-deployment-hkpfs-777898ffcc\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 09:47:47.121: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-deployment-hkpfs-777898ffcc\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 09:47:49.127: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-deployment-hkpfs-777898ffcc\" is progressing."}}, CollisionCount:(*int32)(nil)} + STEP: Getting /status 02/25/23 09:47:51.128 + Feb 25 09:47:51.140: INFO: Deployment test-deployment-hkpfs has Conditions: [{Available True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:49 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-hkpfs-777898ffcc" has successfully progressed.}] + STEP: updating Deployment Status 02/25/23 09:47:51.14 + Feb 25 09:47:51.158: INFO: updatedStatus.Conditions: []v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 49, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 49, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 47, 49, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 47, 41, 0, time.Local), Reason:"NewReplicaSetAvailable", Message:"ReplicaSet \"test-deployment-hkpfs-777898ffcc\" has successfully progressed."}, v1.DeploymentCondition{Type:"StatusUpdate", Status:"True", LastUpdateTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} + STEP: watching for the Deployment status to be updated 02/25/23 09:47:51.158 + Feb 25 09:47:51.162: INFO: Observed &Deployment event: ADDED + Feb 25 09:47:51.163: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-hkpfs-777898ffcc"} + Feb 25 09:47:51.163: INFO: Observed &Deployment event: MODIFIED + Feb 25 09:47:51.163: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-hkpfs-777898ffcc"} + Feb 25 09:47:51.164: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} + Feb 25 09:47:51.164: INFO: Observed &Deployment event: MODIFIED + Feb 25 09:47:51.164: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} + Feb 25 09:47:51.164: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC ReplicaSetUpdated ReplicaSet "test-deployment-hkpfs-777898ffcc" is progressing.} + Feb 25 09:47:51.164: INFO: Observed &Deployment event: MODIFIED + Feb 25 09:47:51.165: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:49 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} + Feb 25 09:47:51.165: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-hkpfs-777898ffcc" has successfully progressed.} + Feb 25 09:47:51.165: INFO: Observed &Deployment event: MODIFIED + Feb 25 09:47:51.165: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:49 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} + Feb 25 09:47:51.165: INFO: Observed Deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-hkpfs-777898ffcc" has successfully progressed.} + Feb 25 09:47:51.165: INFO: Found Deployment test-deployment-hkpfs in namespace deployment-2532 with labels: map[e2e:testing name:httpd] annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} + Feb 25 09:47:51.165: INFO: Deployment test-deployment-hkpfs has an updated status + STEP: patching the Statefulset Status 02/25/23 09:47:51.165 + Feb 25 09:47:51.166: INFO: Patch payload: {"status":{"conditions":[{"type":"StatusPatched","status":"True"}]}} + Feb 25 09:47:51.175: INFO: Patched status conditions: []v1.DeploymentCondition{v1.DeploymentCondition{Type:"StatusPatched", Status:"True", LastUpdateTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"", Message:""}} + STEP: watching for the Deployment status to be patched 02/25/23 09:47:51.175 + Feb 25 09:47:51.178: INFO: Observed &Deployment event: ADDED + Feb 25 09:47:51.178: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-hkpfs-777898ffcc"} + Feb 25 09:47:51.178: INFO: Observed &Deployment event: MODIFIED + Feb 25 09:47:51.178: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-hkpfs-777898ffcc"} + Feb 25 09:47:51.178: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} + Feb 25 09:47:51.179: INFO: Observed &Deployment event: MODIFIED + Feb 25 09:47:51.179: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} + Feb 25 09:47:51.179: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:41 +0000 UTC 2023-02-25 09:47:41 +0000 UTC ReplicaSetUpdated ReplicaSet "test-deployment-hkpfs-777898ffcc" is progressing.} + Feb 25 09:47:51.179: INFO: Observed &Deployment event: MODIFIED + Feb 25 09:47:51.180: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:49 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} + Feb 25 09:47:51.180: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-hkpfs-777898ffcc" has successfully progressed.} + Feb 25 09:47:51.180: INFO: Observed &Deployment event: MODIFIED + Feb 25 09:47:51.180: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:49 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} + Feb 25 09:47:51.180: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2023-02-25 09:47:49 +0000 UTC 2023-02-25 09:47:41 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-hkpfs-777898ffcc" has successfully progressed.} + Feb 25 09:47:51.180: INFO: Observed deployment test-deployment-hkpfs in namespace deployment-2532 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} + Feb 25 09:47:51.180: INFO: Observed &Deployment event: MODIFIED + Feb 25 09:47:51.180: INFO: Found deployment test-deployment-hkpfs in namespace deployment-2532 with labels: map[e2e:testing name:httpd] annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {StatusPatched True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } + Feb 25 09:47:51.180: INFO: Deployment test-deployment-hkpfs has a patched status + [AfterEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:84 + Feb 25 09:47:51.193: INFO: Deployment "test-deployment-hkpfs": + &Deployment{ObjectMeta:{test-deployment-hkpfs deployment-2532 bc5f99c9-0e74-4e89-9947-ffd92c4558f5 8913 1 2023-02-25 09:47:41 +0000 UTC map[e2e:testing name:httpd] map[deployment.kubernetes.io/revision:1] [] [] [{e2e.test Update apps/v1 2023-02-25 09:47:41 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:e2e":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:e2e":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {e2e.test Update apps/v1 2023-02-25 09:47:51 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"StatusPatched\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:status":{},"f:type":{}}}}} status} {kube-controller-manager Update apps/v1 2023-02-25 09:47:51 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{e2e: testing,name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[e2e:testing name:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc001568708 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:StatusPatched,Status:True,Reason:,Message:,LastUpdateTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:0001-01-01 00:00:00 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:FoundNewReplicaSet,Message:Found new replica set "test-deployment-hkpfs-777898ffcc",LastUpdateTime:2023-02-25 09:47:51 +0000 UTC,LastTransitionTime:2023-02-25 09:47:51 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} + + Feb 25 09:47:51.201: INFO: New ReplicaSet "test-deployment-hkpfs-777898ffcc" of Deployment "test-deployment-hkpfs": + &ReplicaSet{ObjectMeta:{test-deployment-hkpfs-777898ffcc deployment-2532 dcf8ee03-84e2-4c62-8743-a4772da4d093 8905 1 2023-02-25 09:47:41 +0000 UTC map[e2e:testing name:httpd pod-template-hash:777898ffcc] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-deployment-hkpfs bc5f99c9-0e74-4e89-9947-ffd92c4558f5 0xc001589237 0xc001589238}] [] [{kube-controller-manager Update apps/v1 2023-02-25 09:47:41 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:e2e":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"bc5f99c9-0e74-4e89-9947-ffd92c4558f5\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:e2e":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:47:49 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{e2e: testing,name: httpd,pod-template-hash: 777898ffcc,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[e2e:testing name:httpd pod-template-hash:777898ffcc] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0015892e8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} + Feb 25 09:47:51.210: INFO: Pod "test-deployment-hkpfs-777898ffcc-djm5x" is available: + &Pod{ObjectMeta:{test-deployment-hkpfs-777898ffcc-djm5x test-deployment-hkpfs-777898ffcc- deployment-2532 c3dbe72b-33e2-4212-8944-f9386f7a5e75 8904 0 2023-02-25 09:47:41 +0000 UTC map[e2e:testing name:httpd pod-template-hash:777898ffcc] map[] [{apps/v1 ReplicaSet test-deployment-hkpfs-777898ffcc dcf8ee03-84e2-4c62-8743-a4772da4d093 0xc001589687 0xc001589688}] [] [{kube-controller-manager Update v1 2023-02-25 09:47:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:e2e":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"dcf8ee03-84e2-4c62-8743-a4772da4d093\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:47:49 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.194\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-rdw9z,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rdw9z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:47:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:47:49 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:47:49 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:47:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:10.233.66.194,StartTime:2023-02-25 09:47:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:47:49 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://b2c62d4cbc88f0f4d735b9954601c7c9a5492b3659bc27ac13bb0657cf797194,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.194,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + [AfterEach] [sig-apps] Deployment test/e2e/framework/framework.go:187 - Dec 14 15:37:19.562: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "sched-pred-6452" for this suite. 12/14/22 15:37:19.57 - [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:83 + Feb 25 09:47:51.210: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "deployment-2532" for this suite. 02/25/23 09:47:51.216 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition - listing custom resource definition objects works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:85 -[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +[sig-node] Downward API + should provide host IP as an env var [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:89 +[BeforeEach] [sig-node] Downward API test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:37:19.586 -Dec 14 15:37:19.587: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename custom-resource-definition 12/14/22 15:37:19.59 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:37:19.619 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:37:19.623 -[It] listing custom resource definition objects works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:85 -Dec 14 15:37:19.628: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 09:47:51.237 +Feb 25 09:47:51.238: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 09:47:51.24 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:51.269 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:51.274 +[It] should provide host IP as an env var [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:89 +STEP: Creating a pod to test downward api env vars 02/25/23 09:47:51.278 +Feb 25 09:47:51.296: INFO: Waiting up to 5m0s for pod "downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f" in namespace "downward-api-4519" to be "Succeeded or Failed" +Feb 25 09:47:51.301: INFO: Pod "downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f": Phase="Pending", Reason="", readiness=false. Elapsed: 5.781003ms +Feb 25 09:47:53.309: INFO: Pod "downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013078658s +Feb 25 09:47:55.308: INFO: Pod "downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012642189s +Feb 25 09:47:57.311: INFO: Pod "downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.015518899s +STEP: Saw pod success 02/25/23 09:47:57.311 +Feb 25 09:47:57.312: INFO: Pod "downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f" satisfied condition "Succeeded or Failed" +Feb 25 09:47:57.320: INFO: Trying to get logs from node diek3dua9gay-3 pod downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f container dapi-container: +STEP: delete the pod 02/25/23 09:47:57.335 +Feb 25 09:47:57.356: INFO: Waiting for pod downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f to disappear +Feb 25 09:47:57.362: INFO: Pod downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f no longer exists +[AfterEach] [sig-node] Downward API test/e2e/framework/framework.go:187 -Dec 14 15:37:25.962: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "custom-resource-definition-6315" for this suite. 12/14/22 15:37:25.982 -{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition listing custom resource definition objects works [Conformance]","completed":8,"skipped":58,"failed":0} +Feb 25 09:47:57.362: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-4519" for this suite. 02/25/23 09:47:57.369 +{"msg":"PASSED [sig-node] Downward API should provide host IP as an env var [NodeConformance] [Conformance]","completed":8,"skipped":268,"failed":0} ------------------------------ -• [SLOW TEST] [6.411 seconds] -[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - Simple CustomResourceDefinition - test/e2e/apimachinery/custom_resource_definition.go:50 - listing custom resource definition objects works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:85 +• [SLOW TEST] [6.143 seconds] +[sig-node] Downward API +test/e2e/common/node/framework.go:23 + should provide host IP as an env var [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:89 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] + [BeforeEach] [sig-node] Downward API test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:37:19.586 - Dec 14 15:37:19.587: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename custom-resource-definition 12/14/22 15:37:19.59 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:37:19.619 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:37:19.623 - [It] listing custom resource definition objects works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:85 - Dec 14 15:37:19.628: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 09:47:51.237 + Feb 25 09:47:51.238: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 09:47:51.24 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:51.269 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:51.274 + [It] should provide host IP as an env var [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:89 + STEP: Creating a pod to test downward api env vars 02/25/23 09:47:51.278 + Feb 25 09:47:51.296: INFO: Waiting up to 5m0s for pod "downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f" in namespace "downward-api-4519" to be "Succeeded or Failed" + Feb 25 09:47:51.301: INFO: Pod "downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f": Phase="Pending", Reason="", readiness=false. Elapsed: 5.781003ms + Feb 25 09:47:53.309: INFO: Pod "downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013078658s + Feb 25 09:47:55.308: INFO: Pod "downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012642189s + Feb 25 09:47:57.311: INFO: Pod "downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.015518899s + STEP: Saw pod success 02/25/23 09:47:57.311 + Feb 25 09:47:57.312: INFO: Pod "downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f" satisfied condition "Succeeded or Failed" + Feb 25 09:47:57.320: INFO: Trying to get logs from node diek3dua9gay-3 pod downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f container dapi-container: + STEP: delete the pod 02/25/23 09:47:57.335 + Feb 25 09:47:57.356: INFO: Waiting for pod downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f to disappear + Feb 25 09:47:57.362: INFO: Pod downward-api-dac057f7-1d93-43b8-b7d3-d7f4eccdfe8f no longer exists + [AfterEach] [sig-node] Downward API test/e2e/framework/framework.go:187 - Dec 14 15:37:25.962: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "custom-resource-definition-6315" for this suite. 12/14/22 15:37:25.982 + Feb 25 09:47:57.362: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-4519" for this suite. 02/25/23 09:47:57.369 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSS +SSSSSSSS ------------------------------ -[sig-storage] EmptyDir wrapper volumes - should not cause race condition when used for configmaps [Serial] [Conformance] - test/e2e/storage/empty_dir_wrapper.go:189 -[BeforeEach] [sig-storage] EmptyDir wrapper volumes +[sig-node] Probing container + with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:68 +[BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:37:26.005 -Dec 14 15:37:26.006: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir-wrapper 12/14/22 15:37:26.009 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:37:26.037 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:37:26.041 -[It] should not cause race condition when used for configmaps [Serial] [Conformance] - test/e2e/storage/empty_dir_wrapper.go:189 -STEP: Creating 50 configmaps 12/14/22 15:37:26.045 -STEP: Creating RC which spawns configmap-volume pods 12/14/22 15:37:26.328 -Dec 14 15:37:26.371: INFO: Pod name wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2: Found 0 pods out of 5 -Dec 14 15:37:31.382: INFO: Pod name wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2: Found 5 pods out of 5 -STEP: Ensuring each pod is running 12/14/22 15:37:31.382 -Dec 14 15:37:31.382: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj" in namespace "emptydir-wrapper-1428" to be "running" -Dec 14 15:37:31.389: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj": Phase="Pending", Reason="", readiness=false. Elapsed: 6.957663ms -Dec 14 15:37:33.398: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015626251s -Dec 14 15:37:35.404: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj": Phase="Pending", Reason="", readiness=false. Elapsed: 4.021906519s -Dec 14 15:37:37.399: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj": Phase="Pending", Reason="", readiness=false. Elapsed: 6.017084502s -Dec 14 15:37:39.399: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj": Phase="Pending", Reason="", readiness=false. Elapsed: 8.016562284s -Dec 14 15:37:41.409: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj": Phase="Running", Reason="", readiness=true. Elapsed: 10.026925263s -Dec 14 15:37:41.410: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj" satisfied condition "running" -Dec 14 15:37:41.410: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-c8xvw" in namespace "emptydir-wrapper-1428" to be "running" -Dec 14 15:37:41.430: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-c8xvw": Phase="Running", Reason="", readiness=true. Elapsed: 19.733746ms -Dec 14 15:37:41.430: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-c8xvw" satisfied condition "running" -Dec 14 15:37:41.430: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-fmrct" in namespace "emptydir-wrapper-1428" to be "running" -Dec 14 15:37:41.439: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-fmrct": Phase="Running", Reason="", readiness=true. Elapsed: 8.83925ms -Dec 14 15:37:41.439: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-fmrct" satisfied condition "running" -Dec 14 15:37:41.440: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-n6tvb" in namespace "emptydir-wrapper-1428" to be "running" -Dec 14 15:37:41.450: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-n6tvb": Phase="Running", Reason="", readiness=true. Elapsed: 10.182719ms -Dec 14 15:37:41.450: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-n6tvb" satisfied condition "running" -Dec 14 15:37:41.450: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-wzbmm" in namespace "emptydir-wrapper-1428" to be "running" -Dec 14 15:37:41.456: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-wzbmm": Phase="Running", Reason="", readiness=true. Elapsed: 5.561296ms -Dec 14 15:37:41.456: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-wzbmm" satisfied condition "running" -STEP: deleting ReplicationController wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2 in namespace emptydir-wrapper-1428, will wait for the garbage collector to delete the pods 12/14/22 15:37:41.456 -Dec 14 15:37:41.526: INFO: Deleting ReplicationController wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2 took: 13.926174ms -Dec 14 15:37:41.627: INFO: Terminating ReplicationController wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2 pods took: 101.332963ms -STEP: Creating RC which spawns configmap-volume pods 12/14/22 15:37:43.938 -Dec 14 15:37:43.966: INFO: Pod name wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7: Found 0 pods out of 5 -Dec 14 15:37:48.981: INFO: Pod name wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7: Found 5 pods out of 5 -STEP: Ensuring each pod is running 12/14/22 15:37:48.981 -Dec 14 15:37:48.982: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd" in namespace "emptydir-wrapper-1428" to be "running" -Dec 14 15:37:48.989: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd": Phase="Pending", Reason="", readiness=false. Elapsed: 7.845316ms -Dec 14 15:37:50.996: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014688599s -Dec 14 15:37:53.000: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd": Phase="Pending", Reason="", readiness=false. Elapsed: 4.018299902s -Dec 14 15:37:54.996: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd": Phase="Pending", Reason="", readiness=false. Elapsed: 6.01452687s -Dec 14 15:37:56.999: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd": Phase="Pending", Reason="", readiness=false. Elapsed: 8.017075565s -Dec 14 15:37:58.998: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd": Phase="Running", Reason="", readiness=true. Elapsed: 10.016293884s -Dec 14 15:37:58.998: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd" satisfied condition "running" -Dec 14 15:37:58.998: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-7bv9m" in namespace "emptydir-wrapper-1428" to be "running" -Dec 14 15:37:59.003: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-7bv9m": Phase="Running", Reason="", readiness=true. Elapsed: 5.223526ms -Dec 14 15:37:59.003: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-7bv9m" satisfied condition "running" -Dec 14 15:37:59.004: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-87hrd" in namespace "emptydir-wrapper-1428" to be "running" -Dec 14 15:37:59.009: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-87hrd": Phase="Running", Reason="", readiness=true. Elapsed: 5.109117ms -Dec 14 15:37:59.009: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-87hrd" satisfied condition "running" -Dec 14 15:37:59.009: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-bl7cc" in namespace "emptydir-wrapper-1428" to be "running" -Dec 14 15:37:59.015: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-bl7cc": Phase="Running", Reason="", readiness=true. Elapsed: 6.005348ms -Dec 14 15:37:59.015: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-bl7cc" satisfied condition "running" -Dec 14 15:37:59.015: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-fjqks" in namespace "emptydir-wrapper-1428" to be "running" -Dec 14 15:37:59.021: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-fjqks": Phase="Running", Reason="", readiness=true. Elapsed: 5.986179ms -Dec 14 15:37:59.021: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-fjqks" satisfied condition "running" -STEP: deleting ReplicationController wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7 in namespace emptydir-wrapper-1428, will wait for the garbage collector to delete the pods 12/14/22 15:37:59.021 -Dec 14 15:37:59.091: INFO: Deleting ReplicationController wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7 took: 12.986193ms -Dec 14 15:37:59.292: INFO: Terminating ReplicationController wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7 pods took: 200.953906ms -STEP: Creating RC which spawns configmap-volume pods 12/14/22 15:38:02.703 -Dec 14 15:38:02.734: INFO: Pod name wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50: Found 0 pods out of 5 -Dec 14 15:38:07.754: INFO: Pod name wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50: Found 5 pods out of 5 -STEP: Ensuring each pod is running 12/14/22 15:38:07.754 -Dec 14 15:38:07.754: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr" in namespace "emptydir-wrapper-1428" to be "running" -Dec 14 15:38:07.761: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr": Phase="Pending", Reason="", readiness=false. Elapsed: 7.030948ms -Dec 14 15:38:09.774: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019625904s -Dec 14 15:38:11.773: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr": Phase="Pending", Reason="", readiness=false. Elapsed: 4.019132209s -Dec 14 15:38:13.770: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr": Phase="Pending", Reason="", readiness=false. Elapsed: 6.015692673s -Dec 14 15:38:15.775: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr": Phase="Pending", Reason="", readiness=false. Elapsed: 8.020727692s -Dec 14 15:38:17.774: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr": Phase="Running", Reason="", readiness=true. Elapsed: 10.019526276s -Dec 14 15:38:17.774: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr" satisfied condition "running" -Dec 14 15:38:17.774: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2z7wq" in namespace "emptydir-wrapper-1428" to be "running" -Dec 14 15:38:17.783: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2z7wq": Phase="Running", Reason="", readiness=true. Elapsed: 8.95436ms -Dec 14 15:38:17.783: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2z7wq" satisfied condition "running" -Dec 14 15:38:17.783: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-4b46d" in namespace "emptydir-wrapper-1428" to be "running" -Dec 14 15:38:17.813: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-4b46d": Phase="Running", Reason="", readiness=true. Elapsed: 29.796204ms -Dec 14 15:38:17.813: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-4b46d" satisfied condition "running" -Dec 14 15:38:17.813: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-bwl5n" in namespace "emptydir-wrapper-1428" to be "running" -Dec 14 15:38:17.823: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-bwl5n": Phase="Running", Reason="", readiness=true. Elapsed: 9.759094ms -Dec 14 15:38:17.823: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-bwl5n" satisfied condition "running" -Dec 14 15:38:17.823: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-cgrlb" in namespace "emptydir-wrapper-1428" to be "running" -Dec 14 15:38:17.831: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-cgrlb": Phase="Running", Reason="", readiness=true. Elapsed: 8.04551ms -Dec 14 15:38:17.831: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-cgrlb" satisfied condition "running" -STEP: deleting ReplicationController wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50 in namespace emptydir-wrapper-1428, will wait for the garbage collector to delete the pods 12/14/22 15:38:17.832 -Dec 14 15:38:17.910: INFO: Deleting ReplicationController wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50 took: 15.22628ms -Dec 14 15:38:18.010: INFO: Terminating ReplicationController wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50 pods took: 100.537793ms -STEP: Cleaning up the configMaps 12/14/22 15:38:21.112 -[AfterEach] [sig-storage] EmptyDir wrapper volumes +STEP: Creating a kubernetes client 02/25/23 09:47:57.384 +Feb 25 09:47:57.384: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-probe 02/25/23 09:47:57.387 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:57.415 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:57.419 +[BeforeEach] [sig-node] Probing container + test/e2e/common/node/container_probe.go:59 +[It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:68 +Feb 25 09:47:57.438: INFO: Waiting up to 5m0s for pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c" in namespace "container-probe-6232" to be "running and ready" +Feb 25 09:47:57.444: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Pending", Reason="", readiness=false. Elapsed: 5.834243ms +Feb 25 09:47:57.444: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Pending, waiting for it to be Running (with Ready = true) +Feb 25 09:47:59.453: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 2.014919312s +Feb 25 09:47:59.453: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) +Feb 25 09:48:01.451: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 4.012909669s +Feb 25 09:48:01.451: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) +Feb 25 09:48:03.451: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 6.012873851s +Feb 25 09:48:03.451: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) +Feb 25 09:48:05.451: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 8.01324923s +Feb 25 09:48:05.451: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) +Feb 25 09:48:07.451: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 10.013239684s +Feb 25 09:48:07.451: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) +Feb 25 09:48:09.453: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 12.014710548s +Feb 25 09:48:09.453: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) +Feb 25 09:48:11.450: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 14.012574076s +Feb 25 09:48:11.450: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) +Feb 25 09:48:13.451: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 16.013397133s +Feb 25 09:48:13.451: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) +Feb 25 09:48:15.452: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 18.013934632s +Feb 25 09:48:15.452: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) +Feb 25 09:48:17.455: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 20.017027964s +Feb 25 09:48:17.455: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) +Feb 25 09:48:19.451: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=true. Elapsed: 22.013508379s +Feb 25 09:48:19.451: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = true) +Feb 25 09:48:19.451: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c" satisfied condition "running and ready" +Feb 25 09:48:19.459: INFO: Container started at 2023-02-25 09:47:58 +0000 UTC, pod became ready at 2023-02-25 09:48:17 +0000 UTC +[AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 -Dec 14 15:38:21.523: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-wrapper-1428" for this suite. 12/14/22 15:38:21.529 -{"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not cause race condition when used for configmaps [Serial] [Conformance]","completed":9,"skipped":67,"failed":0} +Feb 25 09:48:19.459: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-6232" for this suite. 02/25/23 09:48:19.466 +{"msg":"PASSED [sig-node] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]","completed":9,"skipped":276,"failed":0} ------------------------------ -• [SLOW TEST] [55.532 seconds] -[sig-storage] EmptyDir wrapper volumes -test/e2e/storage/utils/framework.go:23 - should not cause race condition when used for configmaps [Serial] [Conformance] - test/e2e/storage/empty_dir_wrapper.go:189 +• [SLOW TEST] [22.098 seconds] +[sig-node] Probing container +test/e2e/common/node/framework.go:23 + with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:68 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir wrapper volumes + [BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:37:26.005 - Dec 14 15:37:26.006: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir-wrapper 12/14/22 15:37:26.009 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:37:26.037 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:37:26.041 - [It] should not cause race condition when used for configmaps [Serial] [Conformance] - test/e2e/storage/empty_dir_wrapper.go:189 - STEP: Creating 50 configmaps 12/14/22 15:37:26.045 - STEP: Creating RC which spawns configmap-volume pods 12/14/22 15:37:26.328 - Dec 14 15:37:26.371: INFO: Pod name wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2: Found 0 pods out of 5 - Dec 14 15:37:31.382: INFO: Pod name wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2: Found 5 pods out of 5 - STEP: Ensuring each pod is running 12/14/22 15:37:31.382 - Dec 14 15:37:31.382: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj" in namespace "emptydir-wrapper-1428" to be "running" - Dec 14 15:37:31.389: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj": Phase="Pending", Reason="", readiness=false. Elapsed: 6.957663ms - Dec 14 15:37:33.398: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015626251s - Dec 14 15:37:35.404: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj": Phase="Pending", Reason="", readiness=false. Elapsed: 4.021906519s - Dec 14 15:37:37.399: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj": Phase="Pending", Reason="", readiness=false. Elapsed: 6.017084502s - Dec 14 15:37:39.399: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj": Phase="Pending", Reason="", readiness=false. Elapsed: 8.016562284s - Dec 14 15:37:41.409: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj": Phase="Running", Reason="", readiness=true. Elapsed: 10.026925263s - Dec 14 15:37:41.410: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-25zjj" satisfied condition "running" - Dec 14 15:37:41.410: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-c8xvw" in namespace "emptydir-wrapper-1428" to be "running" - Dec 14 15:37:41.430: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-c8xvw": Phase="Running", Reason="", readiness=true. Elapsed: 19.733746ms - Dec 14 15:37:41.430: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-c8xvw" satisfied condition "running" - Dec 14 15:37:41.430: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-fmrct" in namespace "emptydir-wrapper-1428" to be "running" - Dec 14 15:37:41.439: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-fmrct": Phase="Running", Reason="", readiness=true. Elapsed: 8.83925ms - Dec 14 15:37:41.439: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-fmrct" satisfied condition "running" - Dec 14 15:37:41.440: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-n6tvb" in namespace "emptydir-wrapper-1428" to be "running" - Dec 14 15:37:41.450: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-n6tvb": Phase="Running", Reason="", readiness=true. Elapsed: 10.182719ms - Dec 14 15:37:41.450: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-n6tvb" satisfied condition "running" - Dec 14 15:37:41.450: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-wzbmm" in namespace "emptydir-wrapper-1428" to be "running" - Dec 14 15:37:41.456: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-wzbmm": Phase="Running", Reason="", readiness=true. Elapsed: 5.561296ms - Dec 14 15:37:41.456: INFO: Pod "wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2-wzbmm" satisfied condition "running" - STEP: deleting ReplicationController wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2 in namespace emptydir-wrapper-1428, will wait for the garbage collector to delete the pods 12/14/22 15:37:41.456 - Dec 14 15:37:41.526: INFO: Deleting ReplicationController wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2 took: 13.926174ms - Dec 14 15:37:41.627: INFO: Terminating ReplicationController wrapped-volume-race-12ae3d40-2702-48c9-8e2a-3b212f8717e2 pods took: 101.332963ms - STEP: Creating RC which spawns configmap-volume pods 12/14/22 15:37:43.938 - Dec 14 15:37:43.966: INFO: Pod name wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7: Found 0 pods out of 5 - Dec 14 15:37:48.981: INFO: Pod name wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7: Found 5 pods out of 5 - STEP: Ensuring each pod is running 12/14/22 15:37:48.981 - Dec 14 15:37:48.982: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd" in namespace "emptydir-wrapper-1428" to be "running" - Dec 14 15:37:48.989: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd": Phase="Pending", Reason="", readiness=false. Elapsed: 7.845316ms - Dec 14 15:37:50.996: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014688599s - Dec 14 15:37:53.000: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd": Phase="Pending", Reason="", readiness=false. Elapsed: 4.018299902s - Dec 14 15:37:54.996: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd": Phase="Pending", Reason="", readiness=false. Elapsed: 6.01452687s - Dec 14 15:37:56.999: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd": Phase="Pending", Reason="", readiness=false. Elapsed: 8.017075565s - Dec 14 15:37:58.998: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd": Phase="Running", Reason="", readiness=true. Elapsed: 10.016293884s - Dec 14 15:37:58.998: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-4bwwd" satisfied condition "running" - Dec 14 15:37:58.998: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-7bv9m" in namespace "emptydir-wrapper-1428" to be "running" - Dec 14 15:37:59.003: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-7bv9m": Phase="Running", Reason="", readiness=true. Elapsed: 5.223526ms - Dec 14 15:37:59.003: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-7bv9m" satisfied condition "running" - Dec 14 15:37:59.004: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-87hrd" in namespace "emptydir-wrapper-1428" to be "running" - Dec 14 15:37:59.009: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-87hrd": Phase="Running", Reason="", readiness=true. Elapsed: 5.109117ms - Dec 14 15:37:59.009: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-87hrd" satisfied condition "running" - Dec 14 15:37:59.009: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-bl7cc" in namespace "emptydir-wrapper-1428" to be "running" - Dec 14 15:37:59.015: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-bl7cc": Phase="Running", Reason="", readiness=true. Elapsed: 6.005348ms - Dec 14 15:37:59.015: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-bl7cc" satisfied condition "running" - Dec 14 15:37:59.015: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-fjqks" in namespace "emptydir-wrapper-1428" to be "running" - Dec 14 15:37:59.021: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-fjqks": Phase="Running", Reason="", readiness=true. Elapsed: 5.986179ms - Dec 14 15:37:59.021: INFO: Pod "wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7-fjqks" satisfied condition "running" - STEP: deleting ReplicationController wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7 in namespace emptydir-wrapper-1428, will wait for the garbage collector to delete the pods 12/14/22 15:37:59.021 - Dec 14 15:37:59.091: INFO: Deleting ReplicationController wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7 took: 12.986193ms - Dec 14 15:37:59.292: INFO: Terminating ReplicationController wrapped-volume-race-6b1b280f-bc74-4d60-af71-9911300ef4a7 pods took: 200.953906ms - STEP: Creating RC which spawns configmap-volume pods 12/14/22 15:38:02.703 - Dec 14 15:38:02.734: INFO: Pod name wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50: Found 0 pods out of 5 - Dec 14 15:38:07.754: INFO: Pod name wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50: Found 5 pods out of 5 - STEP: Ensuring each pod is running 12/14/22 15:38:07.754 - Dec 14 15:38:07.754: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr" in namespace "emptydir-wrapper-1428" to be "running" - Dec 14 15:38:07.761: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr": Phase="Pending", Reason="", readiness=false. Elapsed: 7.030948ms - Dec 14 15:38:09.774: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019625904s - Dec 14 15:38:11.773: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr": Phase="Pending", Reason="", readiness=false. Elapsed: 4.019132209s - Dec 14 15:38:13.770: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr": Phase="Pending", Reason="", readiness=false. Elapsed: 6.015692673s - Dec 14 15:38:15.775: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr": Phase="Pending", Reason="", readiness=false. Elapsed: 8.020727692s - Dec 14 15:38:17.774: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr": Phase="Running", Reason="", readiness=true. Elapsed: 10.019526276s - Dec 14 15:38:17.774: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2f9xr" satisfied condition "running" - Dec 14 15:38:17.774: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2z7wq" in namespace "emptydir-wrapper-1428" to be "running" - Dec 14 15:38:17.783: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2z7wq": Phase="Running", Reason="", readiness=true. Elapsed: 8.95436ms - Dec 14 15:38:17.783: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-2z7wq" satisfied condition "running" - Dec 14 15:38:17.783: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-4b46d" in namespace "emptydir-wrapper-1428" to be "running" - Dec 14 15:38:17.813: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-4b46d": Phase="Running", Reason="", readiness=true. Elapsed: 29.796204ms - Dec 14 15:38:17.813: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-4b46d" satisfied condition "running" - Dec 14 15:38:17.813: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-bwl5n" in namespace "emptydir-wrapper-1428" to be "running" - Dec 14 15:38:17.823: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-bwl5n": Phase="Running", Reason="", readiness=true. Elapsed: 9.759094ms - Dec 14 15:38:17.823: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-bwl5n" satisfied condition "running" - Dec 14 15:38:17.823: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-cgrlb" in namespace "emptydir-wrapper-1428" to be "running" - Dec 14 15:38:17.831: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-cgrlb": Phase="Running", Reason="", readiness=true. Elapsed: 8.04551ms - Dec 14 15:38:17.831: INFO: Pod "wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50-cgrlb" satisfied condition "running" - STEP: deleting ReplicationController wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50 in namespace emptydir-wrapper-1428, will wait for the garbage collector to delete the pods 12/14/22 15:38:17.832 - Dec 14 15:38:17.910: INFO: Deleting ReplicationController wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50 took: 15.22628ms - Dec 14 15:38:18.010: INFO: Terminating ReplicationController wrapped-volume-race-eb129f39-2df8-411c-8c48-4b74bd47ad50 pods took: 100.537793ms - STEP: Cleaning up the configMaps 12/14/22 15:38:21.112 - [AfterEach] [sig-storage] EmptyDir wrapper volumes + STEP: Creating a kubernetes client 02/25/23 09:47:57.384 + Feb 25 09:47:57.384: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-probe 02/25/23 09:47:57.387 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:47:57.415 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:47:57.419 + [BeforeEach] [sig-node] Probing container + test/e2e/common/node/container_probe.go:59 + [It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:68 + Feb 25 09:47:57.438: INFO: Waiting up to 5m0s for pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c" in namespace "container-probe-6232" to be "running and ready" + Feb 25 09:47:57.444: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Pending", Reason="", readiness=false. Elapsed: 5.834243ms + Feb 25 09:47:57.444: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Pending, waiting for it to be Running (with Ready = true) + Feb 25 09:47:59.453: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 2.014919312s + Feb 25 09:47:59.453: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) + Feb 25 09:48:01.451: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 4.012909669s + Feb 25 09:48:01.451: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) + Feb 25 09:48:03.451: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 6.012873851s + Feb 25 09:48:03.451: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) + Feb 25 09:48:05.451: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 8.01324923s + Feb 25 09:48:05.451: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) + Feb 25 09:48:07.451: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 10.013239684s + Feb 25 09:48:07.451: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) + Feb 25 09:48:09.453: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 12.014710548s + Feb 25 09:48:09.453: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) + Feb 25 09:48:11.450: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 14.012574076s + Feb 25 09:48:11.450: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) + Feb 25 09:48:13.451: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 16.013397133s + Feb 25 09:48:13.451: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) + Feb 25 09:48:15.452: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 18.013934632s + Feb 25 09:48:15.452: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) + Feb 25 09:48:17.455: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=false. Elapsed: 20.017027964s + Feb 25 09:48:17.455: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = false) + Feb 25 09:48:19.451: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c": Phase="Running", Reason="", readiness=true. Elapsed: 22.013508379s + Feb 25 09:48:19.451: INFO: The phase of Pod test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c is Running (Ready = true) + Feb 25 09:48:19.451: INFO: Pod "test-webserver-0680aa14-66ab-4a88-8b01-4b9af75fef2c" satisfied condition "running and ready" + Feb 25 09:48:19.459: INFO: Container started at 2023-02-25 09:47:58 +0000 UTC, pod became ready at 2023-02-25 09:48:17 +0000 UTC + [AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 - Dec 14 15:38:21.523: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-wrapper-1428" for this suite. 12/14/22 15:38:21.529 + Feb 25 09:48:19.459: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-probe-6232" for this suite. 02/25/23 09:48:19.466 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Projected downwardAPI - should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:260 -[BeforeEach] [sig-storage] Projected downwardAPI +[sig-storage] Projected secret + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:77 +[BeforeEach] [sig-storage] Projected secret test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:38:21.543 -Dec 14 15:38:21.544: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 15:38:21.547 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:38:21.576 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:38:21.579 -[BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 -[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:260 -STEP: Creating a pod to test downward API volume plugin 12/14/22 15:38:21.582 -Dec 14 15:38:21.594: INFO: Waiting up to 5m0s for pod "downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5" in namespace "projected-9220" to be "Succeeded or Failed" -Dec 14 15:38:21.599: INFO: Pod "downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5": Phase="Pending", Reason="", readiness=false. Elapsed: 4.911993ms -Dec 14 15:38:23.608: INFO: Pod "downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013253413s -Dec 14 15:38:25.610: INFO: Pod "downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015185415s -STEP: Saw pod success 12/14/22 15:38:25.61 -Dec 14 15:38:25.610: INFO: Pod "downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5" satisfied condition "Succeeded or Failed" -Dec 14 15:38:25.616: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5 container client-container: -STEP: delete the pod 12/14/22 15:38:25.63 -Dec 14 15:38:25.650: INFO: Waiting for pod downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5 to disappear -Dec 14 15:38:25.656: INFO: Pod downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5 no longer exists -[AfterEach] [sig-storage] Projected downwardAPI +STEP: Creating a kubernetes client 02/25/23 09:48:19.486 +Feb 25 09:48:19.486: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 09:48:19.49 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:48:19.532 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:48:19.537 +[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:77 +STEP: Creating projection with secret that has name projected-secret-test-map-7a14271d-2f89-4d52-bd08-7ca8c4277478 02/25/23 09:48:19.54 +STEP: Creating a pod to test consume secrets 02/25/23 09:48:19.549 +Feb 25 09:48:19.564: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606" in namespace "projected-2891" to be "Succeeded or Failed" +Feb 25 09:48:19.569: INFO: Pod "pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606": Phase="Pending", Reason="", readiness=false. Elapsed: 5.30486ms +Feb 25 09:48:21.583: INFO: Pod "pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019381919s +Feb 25 09:48:23.575: INFO: Pod "pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011867256s +STEP: Saw pod success 02/25/23 09:48:23.576 +Feb 25 09:48:23.576: INFO: Pod "pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606" satisfied condition "Succeeded or Failed" +Feb 25 09:48:23.582: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606 container projected-secret-volume-test: +STEP: delete the pod 02/25/23 09:48:23.595 +Feb 25 09:48:23.623: INFO: Waiting for pod pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606 to disappear +Feb 25 09:48:23.629: INFO: Pod pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606 no longer exists +[AfterEach] [sig-storage] Projected secret test/e2e/framework/framework.go:187 -Dec 14 15:38:25.657: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-9220" for this suite. 12/14/22 15:38:25.664 -{"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","completed":10,"skipped":90,"failed":0} +Feb 25 09:48:23.630: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-2891" for this suite. 02/25/23 09:48:23.639 +{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","completed":10,"skipped":292,"failed":0} ------------------------------ -• [4.129 seconds] -[sig-storage] Projected downwardAPI +• [4.169 seconds] +[sig-storage] Projected secret test/e2e/common/storage/framework.go:23 - should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:260 + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:77 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected downwardAPI + [BeforeEach] [sig-storage] Projected secret test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:38:21.543 - Dec 14 15:38:21.544: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 15:38:21.547 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:38:21.576 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:38:21.579 - [BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 - [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:260 - STEP: Creating a pod to test downward API volume plugin 12/14/22 15:38:21.582 - Dec 14 15:38:21.594: INFO: Waiting up to 5m0s for pod "downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5" in namespace "projected-9220" to be "Succeeded or Failed" - Dec 14 15:38:21.599: INFO: Pod "downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5": Phase="Pending", Reason="", readiness=false. Elapsed: 4.911993ms - Dec 14 15:38:23.608: INFO: Pod "downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013253413s - Dec 14 15:38:25.610: INFO: Pod "downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015185415s - STEP: Saw pod success 12/14/22 15:38:25.61 - Dec 14 15:38:25.610: INFO: Pod "downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5" satisfied condition "Succeeded or Failed" - Dec 14 15:38:25.616: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5 container client-container: - STEP: delete the pod 12/14/22 15:38:25.63 - Dec 14 15:38:25.650: INFO: Waiting for pod downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5 to disappear - Dec 14 15:38:25.656: INFO: Pod downwardapi-volume-4e0b7591-accf-486c-b535-51fa1fd0c6d5 no longer exists - [AfterEach] [sig-storage] Projected downwardAPI + STEP: Creating a kubernetes client 02/25/23 09:48:19.486 + Feb 25 09:48:19.486: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 09:48:19.49 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:48:19.532 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:48:19.537 + [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:77 + STEP: Creating projection with secret that has name projected-secret-test-map-7a14271d-2f89-4d52-bd08-7ca8c4277478 02/25/23 09:48:19.54 + STEP: Creating a pod to test consume secrets 02/25/23 09:48:19.549 + Feb 25 09:48:19.564: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606" in namespace "projected-2891" to be "Succeeded or Failed" + Feb 25 09:48:19.569: INFO: Pod "pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606": Phase="Pending", Reason="", readiness=false. Elapsed: 5.30486ms + Feb 25 09:48:21.583: INFO: Pod "pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019381919s + Feb 25 09:48:23.575: INFO: Pod "pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011867256s + STEP: Saw pod success 02/25/23 09:48:23.576 + Feb 25 09:48:23.576: INFO: Pod "pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606" satisfied condition "Succeeded or Failed" + Feb 25 09:48:23.582: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606 container projected-secret-volume-test: + STEP: delete the pod 02/25/23 09:48:23.595 + Feb 25 09:48:23.623: INFO: Waiting for pod pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606 to disappear + Feb 25 09:48:23.629: INFO: Pod pod-projected-secrets-518fe548-e1e8-4f6a-9696-558ba0d66606 no longer exists + [AfterEach] [sig-storage] Projected secret test/e2e/framework/framework.go:187 - Dec 14 15:38:25.657: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-9220" for this suite. 12/14/22 15:38:25.664 + Feb 25 09:48:23.630: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-2891" for this suite. 02/25/23 09:48:23.639 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-cli] Kubectl client Kubectl run pod - should create a pod from an image when restart is Never [Conformance] - test/e2e/kubectl/kubectl.go:1711 -[BeforeEach] [sig-cli] Kubectl client +[sig-node] Pods + should support retrieving logs from the container over websockets [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:617 +[BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:38:25.683 -Dec 14 15:38:25.684: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 15:38:25.685 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:38:25.71 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:38:25.714 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 -[BeforeEach] Kubectl run pod - test/e2e/kubectl/kubectl.go:1698 -[It] should create a pod from an image when restart is Never [Conformance] - test/e2e/kubectl/kubectl.go:1711 -STEP: running the image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 12/14/22 15:38:25.719 -Dec 14 15:38:25.720: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-6334 run e2e-test-httpd-pod --restart=Never --pod-running-timeout=2m0s --image=registry.k8s.io/e2e-test-images/httpd:2.4.38-2' -Dec 14 15:38:26.040: INFO: stderr: "" -Dec 14 15:38:26.041: INFO: stdout: "pod/e2e-test-httpd-pod created\n" -STEP: verifying the pod e2e-test-httpd-pod was created 12/14/22 15:38:26.041 -[AfterEach] Kubectl run pod - test/e2e/kubectl/kubectl.go:1702 -Dec 14 15:38:26.047: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-6334 delete pods e2e-test-httpd-pod' -Dec 14 15:38:57.862: INFO: stderr: "" -Dec 14 15:38:57.862: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" -[AfterEach] [sig-cli] Kubectl client +STEP: Creating a kubernetes client 02/25/23 09:48:23.666 +Feb 25 09:48:23.666: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename pods 02/25/23 09:48:23.669 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:48:23.695 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:48:23.7 +[BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 +[It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:617 +Feb 25 09:48:23.703: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: creating the pod 02/25/23 09:48:23.704 +STEP: submitting the pod to kubernetes 02/25/23 09:48:23.705 +Feb 25 09:48:23.724: INFO: Waiting up to 5m0s for pod "pod-logs-websocket-734a3e67-6448-4bd2-945b-b8838d20550e" in namespace "pods-6473" to be "running and ready" +Feb 25 09:48:23.740: INFO: Pod "pod-logs-websocket-734a3e67-6448-4bd2-945b-b8838d20550e": Phase="Pending", Reason="", readiness=false. Elapsed: 16.022109ms +Feb 25 09:48:23.740: INFO: The phase of Pod pod-logs-websocket-734a3e67-6448-4bd2-945b-b8838d20550e is Pending, waiting for it to be Running (with Ready = true) +Feb 25 09:48:25.749: INFO: Pod "pod-logs-websocket-734a3e67-6448-4bd2-945b-b8838d20550e": Phase="Running", Reason="", readiness=true. Elapsed: 2.024911308s +Feb 25 09:48:25.749: INFO: The phase of Pod pod-logs-websocket-734a3e67-6448-4bd2-945b-b8838d20550e is Running (Ready = true) +Feb 25 09:48:25.749: INFO: Pod "pod-logs-websocket-734a3e67-6448-4bd2-945b-b8838d20550e" satisfied condition "running and ready" +[AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 -Dec 14 15:38:57.862: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-6334" for this suite. 12/14/22 15:38:57.876 -{"msg":"PASSED [sig-cli] Kubectl client Kubectl run pod should create a pod from an image when restart is Never [Conformance]","completed":11,"skipped":123,"failed":0} +Feb 25 09:48:25.797: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-6473" for this suite. 02/25/23 09:48:25.808 +{"msg":"PASSED [sig-node] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance]","completed":11,"skipped":314,"failed":0} ------------------------------ -• [SLOW TEST] [32.205 seconds] -[sig-cli] Kubectl client -test/e2e/kubectl/framework.go:23 - Kubectl run pod - test/e2e/kubectl/kubectl.go:1695 - should create a pod from an image when restart is Never [Conformance] - test/e2e/kubectl/kubectl.go:1711 +• [2.153 seconds] +[sig-node] Pods +test/e2e/common/node/framework.go:23 + should support retrieving logs from the container over websockets [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:617 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-cli] Kubectl client + [BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:38:25.683 - Dec 14 15:38:25.684: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 15:38:25.685 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:38:25.71 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:38:25.714 - [BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 - [BeforeEach] Kubectl run pod - test/e2e/kubectl/kubectl.go:1698 - [It] should create a pod from an image when restart is Never [Conformance] - test/e2e/kubectl/kubectl.go:1711 - STEP: running the image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 12/14/22 15:38:25.719 - Dec 14 15:38:25.720: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-6334 run e2e-test-httpd-pod --restart=Never --pod-running-timeout=2m0s --image=registry.k8s.io/e2e-test-images/httpd:2.4.38-2' - Dec 14 15:38:26.040: INFO: stderr: "" - Dec 14 15:38:26.041: INFO: stdout: "pod/e2e-test-httpd-pod created\n" - STEP: verifying the pod e2e-test-httpd-pod was created 12/14/22 15:38:26.041 - [AfterEach] Kubectl run pod - test/e2e/kubectl/kubectl.go:1702 - Dec 14 15:38:26.047: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-6334 delete pods e2e-test-httpd-pod' - Dec 14 15:38:57.862: INFO: stderr: "" - Dec 14 15:38:57.862: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" - [AfterEach] [sig-cli] Kubectl client + STEP: Creating a kubernetes client 02/25/23 09:48:23.666 + Feb 25 09:48:23.666: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename pods 02/25/23 09:48:23.669 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:48:23.695 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:48:23.7 + [BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 + [It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:617 + Feb 25 09:48:23.703: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: creating the pod 02/25/23 09:48:23.704 + STEP: submitting the pod to kubernetes 02/25/23 09:48:23.705 + Feb 25 09:48:23.724: INFO: Waiting up to 5m0s for pod "pod-logs-websocket-734a3e67-6448-4bd2-945b-b8838d20550e" in namespace "pods-6473" to be "running and ready" + Feb 25 09:48:23.740: INFO: Pod "pod-logs-websocket-734a3e67-6448-4bd2-945b-b8838d20550e": Phase="Pending", Reason="", readiness=false. Elapsed: 16.022109ms + Feb 25 09:48:23.740: INFO: The phase of Pod pod-logs-websocket-734a3e67-6448-4bd2-945b-b8838d20550e is Pending, waiting for it to be Running (with Ready = true) + Feb 25 09:48:25.749: INFO: Pod "pod-logs-websocket-734a3e67-6448-4bd2-945b-b8838d20550e": Phase="Running", Reason="", readiness=true. Elapsed: 2.024911308s + Feb 25 09:48:25.749: INFO: The phase of Pod pod-logs-websocket-734a3e67-6448-4bd2-945b-b8838d20550e is Running (Ready = true) + Feb 25 09:48:25.749: INFO: Pod "pod-logs-websocket-734a3e67-6448-4bd2-945b-b8838d20550e" satisfied condition "running and ready" + [AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 - Dec 14 15:38:57.862: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-6334" for this suite. 12/14/22 15:38:57.876 + Feb 25 09:48:25.797: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "pods-6473" for this suite. 02/25/23 09:48:25.808 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Projected downwardAPI - should provide container's memory limit [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:206 -[BeforeEach] [sig-storage] Projected downwardAPI +[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition + creating/deleting custom resource definition objects works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:58 +[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:38:57.894 -Dec 14 15:38:57.895: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 15:38:57.898 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:38:57.924 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:38:57.929 -[BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 -[It] should provide container's memory limit [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:206 -STEP: Creating a pod to test downward API volume plugin 12/14/22 15:38:57.935 -Dec 14 15:38:57.949: INFO: Waiting up to 5m0s for pod "downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2" in namespace "projected-5283" to be "Succeeded or Failed" -Dec 14 15:38:57.953: INFO: Pod "downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2": Phase="Pending", Reason="", readiness=false. Elapsed: 3.710731ms -Dec 14 15:38:59.962: INFO: Pod "downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012503948s -Dec 14 15:39:01.962: INFO: Pod "downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012892316s -STEP: Saw pod success 12/14/22 15:39:01.962 -Dec 14 15:39:01.963: INFO: Pod "downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2" satisfied condition "Succeeded or Failed" -Dec 14 15:39:01.968: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2 container client-container: -STEP: delete the pod 12/14/22 15:39:01.982 -Dec 14 15:39:01.996: INFO: Waiting for pod downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2 to disappear -Dec 14 15:39:02.000: INFO: Pod downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2 no longer exists -[AfterEach] [sig-storage] Projected downwardAPI +STEP: Creating a kubernetes client 02/25/23 09:48:25.843 +Feb 25 09:48:25.843: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename custom-resource-definition 02/25/23 09:48:25.845 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:48:25.882 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:48:25.887 +[It] creating/deleting custom resource definition objects works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:58 +Feb 25 09:48:25.893: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 15:39:02.000: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-5283" for this suite. 12/14/22 15:39:02.006 -{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance]","completed":12,"skipped":183,"failed":0} +Feb 25 09:48:26.937: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "custom-resource-definition-7603" for this suite. 02/25/23 09:48:26.949 +{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works [Conformance]","completed":12,"skipped":385,"failed":0} ------------------------------ -• [4.117 seconds] -[sig-storage] Projected downwardAPI -test/e2e/common/storage/framework.go:23 - should provide container's memory limit [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:206 +• [1.118 seconds] +[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + Simple CustomResourceDefinition + test/e2e/apimachinery/custom_resource_definition.go:50 + creating/deleting custom resource definition objects works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:58 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected downwardAPI + [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:38:57.894 - Dec 14 15:38:57.895: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 15:38:57.898 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:38:57.924 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:38:57.929 - [BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 - [It] should provide container's memory limit [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:206 - STEP: Creating a pod to test downward API volume plugin 12/14/22 15:38:57.935 - Dec 14 15:38:57.949: INFO: Waiting up to 5m0s for pod "downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2" in namespace "projected-5283" to be "Succeeded or Failed" - Dec 14 15:38:57.953: INFO: Pod "downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2": Phase="Pending", Reason="", readiness=false. Elapsed: 3.710731ms - Dec 14 15:38:59.962: INFO: Pod "downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012503948s - Dec 14 15:39:01.962: INFO: Pod "downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012892316s - STEP: Saw pod success 12/14/22 15:39:01.962 - Dec 14 15:39:01.963: INFO: Pod "downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2" satisfied condition "Succeeded or Failed" - Dec 14 15:39:01.968: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2 container client-container: - STEP: delete the pod 12/14/22 15:39:01.982 - Dec 14 15:39:01.996: INFO: Waiting for pod downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2 to disappear - Dec 14 15:39:02.000: INFO: Pod downwardapi-volume-ff78ce71-7b01-49b9-84f5-62742b03c8d2 no longer exists - [AfterEach] [sig-storage] Projected downwardAPI + STEP: Creating a kubernetes client 02/25/23 09:48:25.843 + Feb 25 09:48:25.843: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename custom-resource-definition 02/25/23 09:48:25.845 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:48:25.882 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:48:25.887 + [It] creating/deleting custom resource definition objects works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:58 + Feb 25 09:48:25.893: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 15:39:02.000: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-5283" for this suite. 12/14/22 15:39:02.006 + Feb 25 09:48:26.937: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "custom-resource-definition-7603" for this suite. 02/25/23 09:48:26.949 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] ResourceQuota - should create a ResourceQuota and capture the life of a service. [Conformance] - test/e2e/apimachinery/resource_quota.go:90 -[BeforeEach] [sig-api-machinery] ResourceQuota +[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] + should have a working scale subresource [Conformance] + test/e2e/apps/statefulset.go:846 +[BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:39:02.02 -Dec 14 15:39:02.020: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename resourcequota 12/14/22 15:39:02.029 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:39:02.057 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:39:02.061 -[It] should create a ResourceQuota and capture the life of a service. [Conformance] - test/e2e/apimachinery/resource_quota.go:90 -STEP: Counting existing ResourceQuota 12/14/22 15:39:02.065 -STEP: Creating a ResourceQuota 12/14/22 15:39:07.072 -STEP: Ensuring resource quota status is calculated 12/14/22 15:39:07.079 -STEP: Creating a Service 12/14/22 15:39:09.089 -STEP: Creating a NodePort Service 12/14/22 15:39:09.127 -STEP: Not allowing a LoadBalancer Service with NodePort to be created that exceeds remaining quota 12/14/22 15:39:09.162 -STEP: Ensuring resource quota status captures service creation 12/14/22 15:39:09.199 -STEP: Deleting Services 12/14/22 15:39:11.209 -STEP: Ensuring resource quota status released usage 12/14/22 15:39:11.259 -[AfterEach] [sig-api-machinery] ResourceQuota +STEP: Creating a kubernetes client 02/25/23 09:48:26.965 +Feb 25 09:48:26.965: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename statefulset 02/25/23 09:48:26.967 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:48:27.007 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:48:27.012 +[BeforeEach] [sig-apps] StatefulSet + test/e2e/apps/statefulset.go:96 +[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:111 +STEP: Creating service test in namespace statefulset-678 02/25/23 09:48:27.025 +[It] should have a working scale subresource [Conformance] + test/e2e/apps/statefulset.go:846 +STEP: Creating statefulset ss in namespace statefulset-678 02/25/23 09:48:27.035 +Feb 25 09:48:27.059: INFO: Found 0 stateful pods, waiting for 1 +Feb 25 09:48:37.070: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true +STEP: getting scale subresource 02/25/23 09:48:37.08 +STEP: updating a scale subresource 02/25/23 09:48:37.085 +STEP: verifying the statefulset Spec.Replicas was modified 02/25/23 09:48:37.095 +STEP: Patch a scale subresource 02/25/23 09:48:37.106 +STEP: verifying the statefulset Spec.Replicas was modified 02/25/23 09:48:37.147 +[AfterEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:122 +Feb 25 09:48:37.197: INFO: Deleting all statefulset in ns statefulset-678 +Feb 25 09:48:37.219: INFO: Scaling statefulset ss to 0 +Feb 25 09:49:27.273: INFO: Waiting for statefulset status.replicas updated to 0 +Feb 25 09:49:27.280: INFO: Deleting statefulset ss +[AfterEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:187 -Dec 14 15:39:13.271: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-955" for this suite. 12/14/22 15:39:13.279 -{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a service. [Conformance]","completed":13,"skipped":206,"failed":0} +Feb 25 09:49:27.303: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "statefulset-678" for this suite. 02/25/23 09:49:27.313 +{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should have a working scale subresource [Conformance]","completed":13,"skipped":399,"failed":0} ------------------------------ -• [SLOW TEST] [11.270 seconds] -[sig-api-machinery] ResourceQuota -test/e2e/apimachinery/framework.go:23 - should create a ResourceQuota and capture the life of a service. [Conformance] - test/e2e/apimachinery/resource_quota.go:90 +• [SLOW TEST] [60.365 seconds] +[sig-apps] StatefulSet +test/e2e/apps/framework.go:23 + Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:101 + should have a working scale subresource [Conformance] + test/e2e/apps/statefulset.go:846 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] ResourceQuota + [BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:39:02.02 - Dec 14 15:39:02.020: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename resourcequota 12/14/22 15:39:02.029 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:39:02.057 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:39:02.061 - [It] should create a ResourceQuota and capture the life of a service. [Conformance] - test/e2e/apimachinery/resource_quota.go:90 - STEP: Counting existing ResourceQuota 12/14/22 15:39:02.065 - STEP: Creating a ResourceQuota 12/14/22 15:39:07.072 - STEP: Ensuring resource quota status is calculated 12/14/22 15:39:07.079 - STEP: Creating a Service 12/14/22 15:39:09.089 - STEP: Creating a NodePort Service 12/14/22 15:39:09.127 - STEP: Not allowing a LoadBalancer Service with NodePort to be created that exceeds remaining quota 12/14/22 15:39:09.162 - STEP: Ensuring resource quota status captures service creation 12/14/22 15:39:09.199 - STEP: Deleting Services 12/14/22 15:39:11.209 - STEP: Ensuring resource quota status released usage 12/14/22 15:39:11.259 - [AfterEach] [sig-api-machinery] ResourceQuota + STEP: Creating a kubernetes client 02/25/23 09:48:26.965 + Feb 25 09:48:26.965: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename statefulset 02/25/23 09:48:26.967 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:48:27.007 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:48:27.012 + [BeforeEach] [sig-apps] StatefulSet + test/e2e/apps/statefulset.go:96 + [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:111 + STEP: Creating service test in namespace statefulset-678 02/25/23 09:48:27.025 + [It] should have a working scale subresource [Conformance] + test/e2e/apps/statefulset.go:846 + STEP: Creating statefulset ss in namespace statefulset-678 02/25/23 09:48:27.035 + Feb 25 09:48:27.059: INFO: Found 0 stateful pods, waiting for 1 + Feb 25 09:48:37.070: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true + STEP: getting scale subresource 02/25/23 09:48:37.08 + STEP: updating a scale subresource 02/25/23 09:48:37.085 + STEP: verifying the statefulset Spec.Replicas was modified 02/25/23 09:48:37.095 + STEP: Patch a scale subresource 02/25/23 09:48:37.106 + STEP: verifying the statefulset Spec.Replicas was modified 02/25/23 09:48:37.147 + [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:122 + Feb 25 09:48:37.197: INFO: Deleting all statefulset in ns statefulset-678 + Feb 25 09:48:37.219: INFO: Scaling statefulset ss to 0 + Feb 25 09:49:27.273: INFO: Waiting for statefulset status.replicas updated to 0 + Feb 25 09:49:27.280: INFO: Deleting statefulset ss + [AfterEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:187 - Dec 14 15:39:13.271: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "resourcequota-955" for this suite. 12/14/22 15:39:13.279 + Feb 25 09:49:27.303: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "statefulset-678" for this suite. 02/25/23 09:49:27.313 << End Captured GinkgoWriter Output ------------------------------ -SSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:136 -[BeforeEach] [sig-storage] EmptyDir volumes +[sig-apps] Deployment + RollingUpdateDeployment should delete old pods and create new ones [Conformance] + test/e2e/apps/deployment.go:105 +[BeforeEach] [sig-apps] Deployment test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:39:13.291 -Dec 14 15:39:13.291: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir 12/14/22 15:39:13.293 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:39:13.323 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:39:13.328 -[It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:136 -STEP: Creating a pod to test emptydir 0666 on tmpfs 12/14/22 15:39:13.332 -Dec 14 15:39:13.347: INFO: Waiting up to 5m0s for pod "pod-06b9e04d-4d89-4426-919c-968a3a593629" in namespace "emptydir-6900" to be "Succeeded or Failed" -Dec 14 15:39:13.353: INFO: Pod "pod-06b9e04d-4d89-4426-919c-968a3a593629": Phase="Pending", Reason="", readiness=false. Elapsed: 5.928004ms -Dec 14 15:39:15.366: INFO: Pod "pod-06b9e04d-4d89-4426-919c-968a3a593629": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018658666s -Dec 14 15:39:17.363: INFO: Pod "pod-06b9e04d-4d89-4426-919c-968a3a593629": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015893265s -STEP: Saw pod success 12/14/22 15:39:17.364 -Dec 14 15:39:17.364: INFO: Pod "pod-06b9e04d-4d89-4426-919c-968a3a593629" satisfied condition "Succeeded or Failed" -Dec 14 15:39:17.372: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-06b9e04d-4d89-4426-919c-968a3a593629 container test-container: -STEP: delete the pod 12/14/22 15:39:17.385 -Dec 14 15:39:17.403: INFO: Waiting for pod pod-06b9e04d-4d89-4426-919c-968a3a593629 to disappear -Dec 14 15:39:17.407: INFO: Pod pod-06b9e04d-4d89-4426-919c-968a3a593629 no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +STEP: Creating a kubernetes client 02/25/23 09:49:27.339 +Feb 25 09:49:27.339: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename deployment 02/25/23 09:49:27.342 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:49:27.372 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:49:27.375 +[BeforeEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:91 +[It] RollingUpdateDeployment should delete old pods and create new ones [Conformance] + test/e2e/apps/deployment.go:105 +Feb 25 09:49:27.378: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted) +Feb 25 09:49:27.395: INFO: Pod name sample-pod: Found 0 pods out of 1 +Feb 25 09:49:32.408: INFO: Pod name sample-pod: Found 1 pods out of 1 +STEP: ensuring each pod is running 02/25/23 09:49:32.408 +Feb 25 09:49:32.409: INFO: Creating deployment "test-rolling-update-deployment" +Feb 25 09:49:32.419: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has +Feb 25 09:49:32.429: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created +Feb 25 09:49:34.448: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected +Feb 25 09:49:34.454: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted) +[AfterEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:84 +Feb 25 09:49:34.476: INFO: Deployment "test-rolling-update-deployment": +&Deployment{ObjectMeta:{test-rolling-update-deployment deployment-4262 ced21d4b-8d61-4a51-9b8a-6f88ae69e2b4 9431 1 2023-02-25 09:49:32 +0000 UTC map[name:sample-pod] map[deployment.kubernetes.io/revision:3546343826724305833] [] [] [{e2e.test Update apps/v1 2023-02-25 09:49:32 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:49:34 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc002f32488 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2023-02-25 09:49:32 +0000 UTC,LastTransitionTime:2023-02-25 09:49:32 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rolling-update-deployment-78f575d8ff" has successfully progressed.,LastUpdateTime:2023-02-25 09:49:34 +0000 UTC,LastTransitionTime:2023-02-25 09:49:32 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} + +Feb 25 09:49:34.483: INFO: New ReplicaSet "test-rolling-update-deployment-78f575d8ff" of Deployment "test-rolling-update-deployment": +&ReplicaSet{ObjectMeta:{test-rolling-update-deployment-78f575d8ff deployment-4262 433cf91c-5f76-47cc-887e-0fd749313703 9420 1 2023-02-25 09:49:32 +0000 UTC map[name:sample-pod pod-template-hash:78f575d8ff] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305833] [{apps/v1 Deployment test-rolling-update-deployment ced21d4b-8d61-4a51-9b8a-6f88ae69e2b4 0xc002f32947 0xc002f32948}] [] [{kube-controller-manager Update apps/v1 2023-02-25 09:49:32 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"ced21d4b-8d61-4a51-9b8a-6f88ae69e2b4\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:49:34 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 78f575d8ff,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod-template-hash:78f575d8ff] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc002f329f8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} +Feb 25 09:49:34.483: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment": +Feb 25 09:49:34.483: INFO: &ReplicaSet{ObjectMeta:{test-rolling-update-controller deployment-4262 5beda867-affd-484c-81e7-f50f413dd2bb 9430 2 2023-02-25 09:49:27 +0000 UTC map[name:sample-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305832] [{apps/v1 Deployment test-rolling-update-deployment ced21d4b-8d61-4a51-9b8a-6f88ae69e2b4 0xc002f3281f 0xc002f32830}] [] [{e2e.test Update apps/v1 2023-02-25 09:49:27 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:49:34 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"ced21d4b-8d61-4a51-9b8a-6f88ae69e2b4\"}":{}}},"f:spec":{"f:replicas":{}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:49:34 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc002f328e8 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} +Feb 25 09:49:34.490: INFO: Pod "test-rolling-update-deployment-78f575d8ff-995nr" is available: +&Pod{ObjectMeta:{test-rolling-update-deployment-78f575d8ff-995nr test-rolling-update-deployment-78f575d8ff- deployment-4262 6c99be92-b36c-4967-8dd3-776b1cfc3cfc 9419 0 2023-02-25 09:49:32 +0000 UTC map[name:sample-pod pod-template-hash:78f575d8ff] map[] [{apps/v1 ReplicaSet test-rolling-update-deployment-78f575d8ff 433cf91c-5f76-47cc-887e-0fd749313703 0xc002f32e47 0xc002f32e48}] [] [{kube-controller-manager Update v1 2023-02-25 09:49:32 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"433cf91c-5f76-47cc-887e-0fd749313703\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:49:34 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.189\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-zsdtp,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zsdtp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:49:32 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:49:34 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:49:34 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:49:32 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:10.233.66.189,StartTime:2023-02-25 09:49:32 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:49:33 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,ImageID:registry.k8s.io/e2e-test-images/agnhost@sha256:37dd9ad84b34913b4bc57e00b0a28e8b00f18ca6a5ecec69461aa57402e5c787,ContainerID:cri-o://35be03e0261cc935fa7f9356bf98e400c9ae8034151ffc87890897ecabf4cc1c,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.189,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +[AfterEach] [sig-apps] Deployment test/e2e/framework/framework.go:187 -Dec 14 15:39:17.408: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-6900" for this suite. 12/14/22 15:39:17.414 -{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","completed":14,"skipped":212,"failed":0} +Feb 25 09:49:34.490: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "deployment-4262" for this suite. 02/25/23 09:49:34.501 +{"msg":"PASSED [sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance]","completed":14,"skipped":428,"failed":0} ------------------------------ -• [4.135 seconds] -[sig-storage] EmptyDir volumes -test/e2e/common/storage/framework.go:23 - should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:136 +• [SLOW TEST] [7.175 seconds] +[sig-apps] Deployment +test/e2e/apps/framework.go:23 + RollingUpdateDeployment should delete old pods and create new ones [Conformance] + test/e2e/apps/deployment.go:105 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir volumes + [BeforeEach] [sig-apps] Deployment test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:39:13.291 - Dec 14 15:39:13.291: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir 12/14/22 15:39:13.293 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:39:13.323 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:39:13.328 - [It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:136 - STEP: Creating a pod to test emptydir 0666 on tmpfs 12/14/22 15:39:13.332 - Dec 14 15:39:13.347: INFO: Waiting up to 5m0s for pod "pod-06b9e04d-4d89-4426-919c-968a3a593629" in namespace "emptydir-6900" to be "Succeeded or Failed" - Dec 14 15:39:13.353: INFO: Pod "pod-06b9e04d-4d89-4426-919c-968a3a593629": Phase="Pending", Reason="", readiness=false. Elapsed: 5.928004ms - Dec 14 15:39:15.366: INFO: Pod "pod-06b9e04d-4d89-4426-919c-968a3a593629": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018658666s - Dec 14 15:39:17.363: INFO: Pod "pod-06b9e04d-4d89-4426-919c-968a3a593629": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015893265s - STEP: Saw pod success 12/14/22 15:39:17.364 - Dec 14 15:39:17.364: INFO: Pod "pod-06b9e04d-4d89-4426-919c-968a3a593629" satisfied condition "Succeeded or Failed" - Dec 14 15:39:17.372: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-06b9e04d-4d89-4426-919c-968a3a593629 container test-container: - STEP: delete the pod 12/14/22 15:39:17.385 - Dec 14 15:39:17.403: INFO: Waiting for pod pod-06b9e04d-4d89-4426-919c-968a3a593629 to disappear - Dec 14 15:39:17.407: INFO: Pod pod-06b9e04d-4d89-4426-919c-968a3a593629 no longer exists - [AfterEach] [sig-storage] EmptyDir volumes + STEP: Creating a kubernetes client 02/25/23 09:49:27.339 + Feb 25 09:49:27.339: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename deployment 02/25/23 09:49:27.342 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:49:27.372 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:49:27.375 + [BeforeEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:91 + [It] RollingUpdateDeployment should delete old pods and create new ones [Conformance] + test/e2e/apps/deployment.go:105 + Feb 25 09:49:27.378: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted) + Feb 25 09:49:27.395: INFO: Pod name sample-pod: Found 0 pods out of 1 + Feb 25 09:49:32.408: INFO: Pod name sample-pod: Found 1 pods out of 1 + STEP: ensuring each pod is running 02/25/23 09:49:32.408 + Feb 25 09:49:32.409: INFO: Creating deployment "test-rolling-update-deployment" + Feb 25 09:49:32.419: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has + Feb 25 09:49:32.429: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created + Feb 25 09:49:34.448: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected + Feb 25 09:49:34.454: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted) + [AfterEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:84 + Feb 25 09:49:34.476: INFO: Deployment "test-rolling-update-deployment": + &Deployment{ObjectMeta:{test-rolling-update-deployment deployment-4262 ced21d4b-8d61-4a51-9b8a-6f88ae69e2b4 9431 1 2023-02-25 09:49:32 +0000 UTC map[name:sample-pod] map[deployment.kubernetes.io/revision:3546343826724305833] [] [] [{e2e.test Update apps/v1 2023-02-25 09:49:32 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:49:34 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc002f32488 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2023-02-25 09:49:32 +0000 UTC,LastTransitionTime:2023-02-25 09:49:32 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rolling-update-deployment-78f575d8ff" has successfully progressed.,LastUpdateTime:2023-02-25 09:49:34 +0000 UTC,LastTransitionTime:2023-02-25 09:49:32 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} + + Feb 25 09:49:34.483: INFO: New ReplicaSet "test-rolling-update-deployment-78f575d8ff" of Deployment "test-rolling-update-deployment": + &ReplicaSet{ObjectMeta:{test-rolling-update-deployment-78f575d8ff deployment-4262 433cf91c-5f76-47cc-887e-0fd749313703 9420 1 2023-02-25 09:49:32 +0000 UTC map[name:sample-pod pod-template-hash:78f575d8ff] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305833] [{apps/v1 Deployment test-rolling-update-deployment ced21d4b-8d61-4a51-9b8a-6f88ae69e2b4 0xc002f32947 0xc002f32948}] [] [{kube-controller-manager Update apps/v1 2023-02-25 09:49:32 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"ced21d4b-8d61-4a51-9b8a-6f88ae69e2b4\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:49:34 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 78f575d8ff,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod-template-hash:78f575d8ff] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc002f329f8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} + Feb 25 09:49:34.483: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment": + Feb 25 09:49:34.483: INFO: &ReplicaSet{ObjectMeta:{test-rolling-update-controller deployment-4262 5beda867-affd-484c-81e7-f50f413dd2bb 9430 2 2023-02-25 09:49:27 +0000 UTC map[name:sample-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305832] [{apps/v1 Deployment test-rolling-update-deployment ced21d4b-8d61-4a51-9b8a-6f88ae69e2b4 0xc002f3281f 0xc002f32830}] [] [{e2e.test Update apps/v1 2023-02-25 09:49:27 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:49:34 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"ced21d4b-8d61-4a51-9b8a-6f88ae69e2b4\"}":{}}},"f:spec":{"f:replicas":{}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:49:34 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc002f328e8 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} + Feb 25 09:49:34.490: INFO: Pod "test-rolling-update-deployment-78f575d8ff-995nr" is available: + &Pod{ObjectMeta:{test-rolling-update-deployment-78f575d8ff-995nr test-rolling-update-deployment-78f575d8ff- deployment-4262 6c99be92-b36c-4967-8dd3-776b1cfc3cfc 9419 0 2023-02-25 09:49:32 +0000 UTC map[name:sample-pod pod-template-hash:78f575d8ff] map[] [{apps/v1 ReplicaSet test-rolling-update-deployment-78f575d8ff 433cf91c-5f76-47cc-887e-0fd749313703 0xc002f32e47 0xc002f32e48}] [] [{kube-controller-manager Update v1 2023-02-25 09:49:32 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"433cf91c-5f76-47cc-887e-0fd749313703\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:49:34 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.189\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-zsdtp,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zsdtp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:49:32 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:49:34 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:49:34 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:49:32 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:10.233.66.189,StartTime:2023-02-25 09:49:32 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:49:33 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,ImageID:registry.k8s.io/e2e-test-images/agnhost@sha256:37dd9ad84b34913b4bc57e00b0a28e8b00f18ca6a5ecec69461aa57402e5c787,ContainerID:cri-o://35be03e0261cc935fa7f9356bf98e400c9ae8034151ffc87890897ecabf4cc1c,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.189,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + [AfterEach] [sig-apps] Deployment test/e2e/framework/framework.go:187 - Dec 14 15:39:17.408: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-6900" for this suite. 12/14/22 15:39:17.414 + Feb 25 09:49:34.490: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "deployment-4262" for this suite. 02/25/23 09:49:34.501 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSS +SSSSSSSSSSS ------------------------------ -[sig-apps] CronJob - should not schedule jobs when suspended [Slow] [Conformance] - test/e2e/apps/cronjob.go:96 -[BeforeEach] [sig-apps] CronJob +[sig-network] Services + should delete a collection of services [Conformance] + test/e2e/network/service.go:3641 +[BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:39:17.429 -Dec 14 15:39:17.429: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename cronjob 12/14/22 15:39:17.442 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:39:17.47 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:39:17.474 -[It] should not schedule jobs when suspended [Slow] [Conformance] - test/e2e/apps/cronjob.go:96 -STEP: Creating a suspended cronjob 12/14/22 15:39:17.479 -STEP: Ensuring no jobs are scheduled 12/14/22 15:39:17.491 -STEP: Ensuring no job exists by listing jobs explicitly 12/14/22 15:44:17.509 -STEP: Removing cronjob 12/14/22 15:44:17.516 -[AfterEach] [sig-apps] CronJob +STEP: Creating a kubernetes client 02/25/23 09:49:34.522 +Feb 25 09:49:34.522: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 09:49:34.524 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:49:34.556 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:49:34.56 +[BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 +[It] should delete a collection of services [Conformance] + test/e2e/network/service.go:3641 +STEP: creating a collection of services 02/25/23 09:49:34.565 +Feb 25 09:49:34.566: INFO: Creating e2e-svc-a-pr72s +Feb 25 09:49:34.591: INFO: Creating e2e-svc-b-gd9wq +Feb 25 09:49:34.606: INFO: Creating e2e-svc-c-9xsqp +STEP: deleting service collection 02/25/23 09:49:34.632 +Feb 25 09:49:34.682: INFO: Collection of services has been deleted +[AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 15:44:17.528: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "cronjob-5009" for this suite. 12/14/22 15:44:17.535 -{"msg":"PASSED [sig-apps] CronJob should not schedule jobs when suspended [Slow] [Conformance]","completed":15,"skipped":224,"failed":0} +Feb 25 09:49:34.682: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-9646" for this suite. 02/25/23 09:49:34.691 +[AfterEach] [sig-network] Services + test/e2e/network/service.go:762 +{"msg":"PASSED [sig-network] Services should delete a collection of services [Conformance]","completed":15,"skipped":439,"failed":0} ------------------------------ -• [SLOW TEST] [300.119 seconds] -[sig-apps] CronJob -test/e2e/apps/framework.go:23 - should not schedule jobs when suspended [Slow] [Conformance] - test/e2e/apps/cronjob.go:96 +• [0.182 seconds] +[sig-network] Services +test/e2e/network/common/framework.go:23 + should delete a collection of services [Conformance] + test/e2e/network/service.go:3641 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] CronJob + [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:39:17.429 - Dec 14 15:39:17.429: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename cronjob 12/14/22 15:39:17.442 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:39:17.47 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:39:17.474 - [It] should not schedule jobs when suspended [Slow] [Conformance] - test/e2e/apps/cronjob.go:96 - STEP: Creating a suspended cronjob 12/14/22 15:39:17.479 - STEP: Ensuring no jobs are scheduled 12/14/22 15:39:17.491 - STEP: Ensuring no job exists by listing jobs explicitly 12/14/22 15:44:17.509 - STEP: Removing cronjob 12/14/22 15:44:17.516 - [AfterEach] [sig-apps] CronJob + STEP: Creating a kubernetes client 02/25/23 09:49:34.522 + Feb 25 09:49:34.522: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 09:49:34.524 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:49:34.556 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:49:34.56 + [BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 + [It] should delete a collection of services [Conformance] + test/e2e/network/service.go:3641 + STEP: creating a collection of services 02/25/23 09:49:34.565 + Feb 25 09:49:34.566: INFO: Creating e2e-svc-a-pr72s + Feb 25 09:49:34.591: INFO: Creating e2e-svc-b-gd9wq + Feb 25 09:49:34.606: INFO: Creating e2e-svc-c-9xsqp + STEP: deleting service collection 02/25/23 09:49:34.632 + Feb 25 09:49:34.682: INFO: Collection of services has been deleted + [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 15:44:17.528: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "cronjob-5009" for this suite. 12/14/22 15:44:17.535 + Feb 25 09:49:34.682: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-9646" for this suite. 02/25/23 09:49:34.691 + [AfterEach] [sig-network] Services + test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Downward API volume - should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:67 -[BeforeEach] [sig-storage] Downward API volume +[sig-storage] Projected downwardAPI + should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:248 +[BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:44:17.555 -Dec 14 15:44:17.555: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 15:44:17.558 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:44:17.589 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:44:17.594 -[BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 -[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:67 -STEP: Creating a pod to test downward API volume plugin 12/14/22 15:44:17.598 -Dec 14 15:44:17.627: INFO: Waiting up to 5m0s for pod "downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905" in namespace "downward-api-7080" to be "Succeeded or Failed" -Dec 14 15:44:17.644: INFO: Pod "downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905": Phase="Pending", Reason="", readiness=false. Elapsed: 16.904027ms -Dec 14 15:44:19.655: INFO: Pod "downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027280108s -Dec 14 15:44:21.655: INFO: Pod "downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.027670367s -STEP: Saw pod success 12/14/22 15:44:21.655 -Dec 14 15:44:21.655: INFO: Pod "downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905" satisfied condition "Succeeded or Failed" -Dec 14 15:44:21.664: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905 container client-container: -STEP: delete the pod 12/14/22 15:44:21.702 -Dec 14 15:44:21.720: INFO: Waiting for pod downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905 to disappear -Dec 14 15:44:21.725: INFO: Pod downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905 no longer exists -[AfterEach] [sig-storage] Downward API volume +STEP: Creating a kubernetes client 02/25/23 09:49:34.729 +Feb 25 09:49:34.729: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 09:49:34.732 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:49:34.76 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:49:34.764 +[BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 +[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:248 +STEP: Creating a pod to test downward API volume plugin 02/25/23 09:49:34.768 +Feb 25 09:49:34.779: INFO: Waiting up to 5m0s for pod "downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf" in namespace "projected-1225" to be "Succeeded or Failed" +Feb 25 09:49:34.784: INFO: Pod "downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf": Phase="Pending", Reason="", readiness=false. Elapsed: 4.884567ms +Feb 25 09:49:36.795: INFO: Pod "downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015784208s +Feb 25 09:49:38.794: INFO: Pod "downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014564453s +STEP: Saw pod success 02/25/23 09:49:38.794 +Feb 25 09:49:38.794: INFO: Pod "downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf" satisfied condition "Succeeded or Failed" +Feb 25 09:49:38.801: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf container client-container: +STEP: delete the pod 02/25/23 09:49:38.816 +Feb 25 09:49:38.839: INFO: Waiting for pod downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf to disappear +Feb 25 09:49:38.846: INFO: Pod downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf no longer exists +[AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 -Dec 14 15:44:21.725: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-7080" for this suite. 12/14/22 15:44:21.731 -{"msg":"PASSED [sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","completed":16,"skipped":239,"failed":0} +Feb 25 09:49:38.847: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-1225" for this suite. 02/25/23 09:49:38.852 +{"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","completed":16,"skipped":474,"failed":0} ------------------------------ -• [4.189 seconds] -[sig-storage] Downward API volume +• [4.136 seconds] +[sig-storage] Projected downwardAPI test/e2e/common/storage/framework.go:23 - should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:67 + should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:248 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Downward API volume + [BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:44:17.555 - Dec 14 15:44:17.555: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 15:44:17.558 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:44:17.589 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:44:17.594 - [BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 - [It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:67 - STEP: Creating a pod to test downward API volume plugin 12/14/22 15:44:17.598 - Dec 14 15:44:17.627: INFO: Waiting up to 5m0s for pod "downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905" in namespace "downward-api-7080" to be "Succeeded or Failed" - Dec 14 15:44:17.644: INFO: Pod "downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905": Phase="Pending", Reason="", readiness=false. Elapsed: 16.904027ms - Dec 14 15:44:19.655: INFO: Pod "downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027280108s - Dec 14 15:44:21.655: INFO: Pod "downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.027670367s - STEP: Saw pod success 12/14/22 15:44:21.655 - Dec 14 15:44:21.655: INFO: Pod "downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905" satisfied condition "Succeeded or Failed" - Dec 14 15:44:21.664: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905 container client-container: - STEP: delete the pod 12/14/22 15:44:21.702 - Dec 14 15:44:21.720: INFO: Waiting for pod downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905 to disappear - Dec 14 15:44:21.725: INFO: Pod downwardapi-volume-276cc4ef-f276-431a-9eab-27bd7b93c905 no longer exists - [AfterEach] [sig-storage] Downward API volume + STEP: Creating a kubernetes client 02/25/23 09:49:34.729 + Feb 25 09:49:34.729: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 09:49:34.732 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:49:34.76 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:49:34.764 + [BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 + [It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:248 + STEP: Creating a pod to test downward API volume plugin 02/25/23 09:49:34.768 + Feb 25 09:49:34.779: INFO: Waiting up to 5m0s for pod "downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf" in namespace "projected-1225" to be "Succeeded or Failed" + Feb 25 09:49:34.784: INFO: Pod "downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf": Phase="Pending", Reason="", readiness=false. Elapsed: 4.884567ms + Feb 25 09:49:36.795: INFO: Pod "downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015784208s + Feb 25 09:49:38.794: INFO: Pod "downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014564453s + STEP: Saw pod success 02/25/23 09:49:38.794 + Feb 25 09:49:38.794: INFO: Pod "downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf" satisfied condition "Succeeded or Failed" + Feb 25 09:49:38.801: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf container client-container: + STEP: delete the pod 02/25/23 09:49:38.816 + Feb 25 09:49:38.839: INFO: Waiting for pod downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf to disappear + Feb 25 09:49:38.846: INFO: Pod downwardapi-volume-8c754c8b-6661-4c0d-b316-3e609173bcbf no longer exists + [AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 - Dec 14 15:44:21.725: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-7080" for this suite. 12/14/22 15:44:21.731 + Feb 25 09:49:38.847: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-1225" for this suite. 02/25/23 09:49:38.852 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Subpath Atomic writer volumes - should support subpaths with downward pod [Conformance] - test/e2e/storage/subpath.go:92 -[BeforeEach] [sig-storage] Subpath +[sig-storage] Projected secret + optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:214 +[BeforeEach] [sig-storage] Projected secret test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:44:21.749 -Dec 14 15:44:21.749: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename subpath 12/14/22 15:44:21.75 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:44:21.785 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:44:21.792 -[BeforeEach] Atomic writer volumes - test/e2e/storage/subpath.go:40 -STEP: Setting up data 12/14/22 15:44:21.799 -[It] should support subpaths with downward pod [Conformance] - test/e2e/storage/subpath.go:92 -STEP: Creating pod pod-subpath-test-downwardapi-xw9w 12/14/22 15:44:21.831 -STEP: Creating a pod to test atomic-volume-subpath 12/14/22 15:44:21.832 -Dec 14 15:44:21.847: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-xw9w" in namespace "subpath-2199" to be "Succeeded or Failed" -Dec 14 15:44:21.851: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Pending", Reason="", readiness=false. Elapsed: 4.316691ms -Dec 14 15:44:23.859: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 2.012245321s -Dec 14 15:44:25.862: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 4.01514897s -Dec 14 15:44:27.862: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 6.015060664s -Dec 14 15:44:29.863: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 8.016384434s -Dec 14 15:44:31.861: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 10.013763156s -Dec 14 15:44:33.863: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 12.015779457s -Dec 14 15:44:35.860: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 14.013621173s -Dec 14 15:44:37.861: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 16.014388551s -Dec 14 15:44:39.863: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 18.015637671s -Dec 14 15:44:41.860: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 20.013380151s -Dec 14 15:44:43.862: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=false. Elapsed: 22.015489512s -Dec 14 15:44:45.861: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.014437921s -STEP: Saw pod success 12/14/22 15:44:45.862 -Dec 14 15:44:45.862: INFO: Pod "pod-subpath-test-downwardapi-xw9w" satisfied condition "Succeeded or Failed" -Dec 14 15:44:45.868: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-subpath-test-downwardapi-xw9w container test-container-subpath-downwardapi-xw9w: -STEP: delete the pod 12/14/22 15:44:45.889 -Dec 14 15:44:45.909: INFO: Waiting for pod pod-subpath-test-downwardapi-xw9w to disappear -Dec 14 15:44:45.913: INFO: Pod pod-subpath-test-downwardapi-xw9w no longer exists -STEP: Deleting pod pod-subpath-test-downwardapi-xw9w 12/14/22 15:44:45.913 -Dec 14 15:44:45.914: INFO: Deleting pod "pod-subpath-test-downwardapi-xw9w" in namespace "subpath-2199" -[AfterEach] [sig-storage] Subpath +STEP: Creating a kubernetes client 02/25/23 09:49:38.874 +Feb 25 09:49:38.874: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 09:49:38.876 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:49:38.904 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:49:38.909 +[It] optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:214 +STEP: Creating secret with name s-test-opt-del-21c6e7af-2719-4100-aff7-23fd124eb620 02/25/23 09:49:38.92 +STEP: Creating secret with name s-test-opt-upd-5ad695b5-09fe-4388-8994-998b1e1f8009 02/25/23 09:49:38.931 +STEP: Creating the pod 02/25/23 09:49:38.939 +Feb 25 09:49:38.956: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071" in namespace "projected-8413" to be "running and ready" +Feb 25 09:49:38.960: INFO: Pod "pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071": Phase="Pending", Reason="", readiness=false. Elapsed: 4.438113ms +Feb 25 09:49:38.960: INFO: The phase of Pod pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 09:49:40.971: INFO: Pod "pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015256631s +Feb 25 09:49:40.972: INFO: The phase of Pod pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 09:49:42.971: INFO: Pod "pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071": Phase="Running", Reason="", readiness=true. Elapsed: 4.015244819s +Feb 25 09:49:42.971: INFO: The phase of Pod pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071 is Running (Ready = true) +Feb 25 09:49:42.972: INFO: Pod "pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071" satisfied condition "running and ready" +STEP: Deleting secret s-test-opt-del-21c6e7af-2719-4100-aff7-23fd124eb620 02/25/23 09:49:43.022 +STEP: Updating secret s-test-opt-upd-5ad695b5-09fe-4388-8994-998b1e1f8009 02/25/23 09:49:43.034 +STEP: Creating secret with name s-test-opt-create-d3cb33b2-cba2-486b-b5f8-8eeafa917b4e 02/25/23 09:49:43.042 +STEP: waiting to observe update in volume 02/25/23 09:49:43.048 +[AfterEach] [sig-storage] Projected secret test/e2e/framework/framework.go:187 -Dec 14 15:44:45.919: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "subpath-2199" for this suite. 12/14/22 15:44:45.925 -{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [Conformance]","completed":17,"skipped":258,"failed":0} +Feb 25 09:50:45.664: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-8413" for this suite. 02/25/23 09:50:45.671 +{"msg":"PASSED [sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance]","completed":17,"skipped":501,"failed":0} ------------------------------ -• [SLOW TEST] [24.193 seconds] -[sig-storage] Subpath -test/e2e/storage/utils/framework.go:23 - Atomic writer volumes - test/e2e/storage/subpath.go:36 - should support subpaths with downward pod [Conformance] - test/e2e/storage/subpath.go:92 +• [SLOW TEST] [66.811 seconds] +[sig-storage] Projected secret +test/e2e/common/storage/framework.go:23 + optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:214 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Subpath + [BeforeEach] [sig-storage] Projected secret test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:44:21.749 - Dec 14 15:44:21.749: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename subpath 12/14/22 15:44:21.75 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:44:21.785 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:44:21.792 - [BeforeEach] Atomic writer volumes - test/e2e/storage/subpath.go:40 - STEP: Setting up data 12/14/22 15:44:21.799 - [It] should support subpaths with downward pod [Conformance] - test/e2e/storage/subpath.go:92 - STEP: Creating pod pod-subpath-test-downwardapi-xw9w 12/14/22 15:44:21.831 - STEP: Creating a pod to test atomic-volume-subpath 12/14/22 15:44:21.832 - Dec 14 15:44:21.847: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-xw9w" in namespace "subpath-2199" to be "Succeeded or Failed" - Dec 14 15:44:21.851: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Pending", Reason="", readiness=false. Elapsed: 4.316691ms - Dec 14 15:44:23.859: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 2.012245321s - Dec 14 15:44:25.862: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 4.01514897s - Dec 14 15:44:27.862: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 6.015060664s - Dec 14 15:44:29.863: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 8.016384434s - Dec 14 15:44:31.861: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 10.013763156s - Dec 14 15:44:33.863: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 12.015779457s - Dec 14 15:44:35.860: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 14.013621173s - Dec 14 15:44:37.861: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 16.014388551s - Dec 14 15:44:39.863: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 18.015637671s - Dec 14 15:44:41.860: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=true. Elapsed: 20.013380151s - Dec 14 15:44:43.862: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Running", Reason="", readiness=false. Elapsed: 22.015489512s - Dec 14 15:44:45.861: INFO: Pod "pod-subpath-test-downwardapi-xw9w": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.014437921s - STEP: Saw pod success 12/14/22 15:44:45.862 - Dec 14 15:44:45.862: INFO: Pod "pod-subpath-test-downwardapi-xw9w" satisfied condition "Succeeded or Failed" - Dec 14 15:44:45.868: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-subpath-test-downwardapi-xw9w container test-container-subpath-downwardapi-xw9w: - STEP: delete the pod 12/14/22 15:44:45.889 - Dec 14 15:44:45.909: INFO: Waiting for pod pod-subpath-test-downwardapi-xw9w to disappear - Dec 14 15:44:45.913: INFO: Pod pod-subpath-test-downwardapi-xw9w no longer exists - STEP: Deleting pod pod-subpath-test-downwardapi-xw9w 12/14/22 15:44:45.913 - Dec 14 15:44:45.914: INFO: Deleting pod "pod-subpath-test-downwardapi-xw9w" in namespace "subpath-2199" - [AfterEach] [sig-storage] Subpath + STEP: Creating a kubernetes client 02/25/23 09:49:38.874 + Feb 25 09:49:38.874: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 09:49:38.876 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:49:38.904 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:49:38.909 + [It] optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:214 + STEP: Creating secret with name s-test-opt-del-21c6e7af-2719-4100-aff7-23fd124eb620 02/25/23 09:49:38.92 + STEP: Creating secret with name s-test-opt-upd-5ad695b5-09fe-4388-8994-998b1e1f8009 02/25/23 09:49:38.931 + STEP: Creating the pod 02/25/23 09:49:38.939 + Feb 25 09:49:38.956: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071" in namespace "projected-8413" to be "running and ready" + Feb 25 09:49:38.960: INFO: Pod "pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071": Phase="Pending", Reason="", readiness=false. Elapsed: 4.438113ms + Feb 25 09:49:38.960: INFO: The phase of Pod pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 09:49:40.971: INFO: Pod "pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015256631s + Feb 25 09:49:40.972: INFO: The phase of Pod pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 09:49:42.971: INFO: Pod "pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071": Phase="Running", Reason="", readiness=true. Elapsed: 4.015244819s + Feb 25 09:49:42.971: INFO: The phase of Pod pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071 is Running (Ready = true) + Feb 25 09:49:42.972: INFO: Pod "pod-projected-secrets-7c6ee6b8-1ff1-4911-8d93-87f0cd1f9071" satisfied condition "running and ready" + STEP: Deleting secret s-test-opt-del-21c6e7af-2719-4100-aff7-23fd124eb620 02/25/23 09:49:43.022 + STEP: Updating secret s-test-opt-upd-5ad695b5-09fe-4388-8994-998b1e1f8009 02/25/23 09:49:43.034 + STEP: Creating secret with name s-test-opt-create-d3cb33b2-cba2-486b-b5f8-8eeafa917b4e 02/25/23 09:49:43.042 + STEP: waiting to observe update in volume 02/25/23 09:49:43.048 + [AfterEach] [sig-storage] Projected secret test/e2e/framework/framework.go:187 - Dec 14 15:44:45.919: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "subpath-2199" for this suite. 12/14/22 15:44:45.925 + Feb 25 09:50:45.664: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-8413" for this suite. 02/25/23 09:50:45.671 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Projected configMap - should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:98 -[BeforeEach] [sig-storage] Projected configMap +[sig-node] Pods + should get a host IP [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:203 +[BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:44:45.945 -Dec 14 15:44:45.946: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 15:44:45.949 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:44:45.981 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:44:45.985 -[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:98 -STEP: Creating configMap with name projected-configmap-test-volume-map-a2798340-72f7-4996-9ffa-06ef88583bfa 12/14/22 15:44:45.989 -STEP: Creating a pod to test consume configMaps 12/14/22 15:44:45.995 -Dec 14 15:44:46.012: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1" in namespace "projected-762" to be "Succeeded or Failed" -Dec 14 15:44:46.016: INFO: Pod "pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1": Phase="Pending", Reason="", readiness=false. Elapsed: 3.777418ms -Dec 14 15:44:48.026: INFO: Pod "pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014151195s -Dec 14 15:44:50.024: INFO: Pod "pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012026087s -STEP: Saw pod success 12/14/22 15:44:50.024 -Dec 14 15:44:50.025: INFO: Pod "pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1" satisfied condition "Succeeded or Failed" -Dec 14 15:44:50.030: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1 container agnhost-container: -STEP: delete the pod 12/14/22 15:44:50.043 -Dec 14 15:44:50.065: INFO: Waiting for pod pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1 to disappear -Dec 14 15:44:50.071: INFO: Pod pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1 no longer exists -[AfterEach] [sig-storage] Projected configMap +STEP: Creating a kubernetes client 02/25/23 09:50:45.691 +Feb 25 09:50:45.691: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename pods 02/25/23 09:50:45.694 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:50:45.717 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:50:45.72 +[BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 +[It] should get a host IP [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:203 +STEP: creating pod 02/25/23 09:50:45.724 +Feb 25 09:50:45.740: INFO: Waiting up to 5m0s for pod "pod-hostip-f26c78f5-20f9-4273-a7e5-cbbb4fc5e963" in namespace "pods-6602" to be "running and ready" +Feb 25 09:50:45.746: INFO: Pod "pod-hostip-f26c78f5-20f9-4273-a7e5-cbbb4fc5e963": Phase="Pending", Reason="", readiness=false. Elapsed: 5.62358ms +Feb 25 09:50:45.746: INFO: The phase of Pod pod-hostip-f26c78f5-20f9-4273-a7e5-cbbb4fc5e963 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 09:50:47.752: INFO: Pod "pod-hostip-f26c78f5-20f9-4273-a7e5-cbbb4fc5e963": Phase="Running", Reason="", readiness=true. Elapsed: 2.011449349s +Feb 25 09:50:47.752: INFO: The phase of Pod pod-hostip-f26c78f5-20f9-4273-a7e5-cbbb4fc5e963 is Running (Ready = true) +Feb 25 09:50:47.752: INFO: Pod "pod-hostip-f26c78f5-20f9-4273-a7e5-cbbb4fc5e963" satisfied condition "running and ready" +Feb 25 09:50:47.763: INFO: Pod pod-hostip-f26c78f5-20f9-4273-a7e5-cbbb4fc5e963 has hostIP: 192.168.121.117 +[AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 -Dec 14 15:44:50.071: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-762" for this suite. 12/14/22 15:44:50.077 -{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","completed":18,"skipped":280,"failed":0} +Feb 25 09:50:47.764: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-6602" for this suite. 02/25/23 09:50:47.772 +{"msg":"PASSED [sig-node] Pods should get a host IP [NodeConformance] [Conformance]","completed":18,"skipped":517,"failed":0} ------------------------------ -• [4.145 seconds] -[sig-storage] Projected configMap -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:98 +• [2.095 seconds] +[sig-node] Pods +test/e2e/common/node/framework.go:23 + should get a host IP [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:203 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected configMap + [BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:44:45.945 - Dec 14 15:44:45.946: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 15:44:45.949 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:44:45.981 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:44:45.985 - [It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:98 - STEP: Creating configMap with name projected-configmap-test-volume-map-a2798340-72f7-4996-9ffa-06ef88583bfa 12/14/22 15:44:45.989 - STEP: Creating a pod to test consume configMaps 12/14/22 15:44:45.995 - Dec 14 15:44:46.012: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1" in namespace "projected-762" to be "Succeeded or Failed" - Dec 14 15:44:46.016: INFO: Pod "pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1": Phase="Pending", Reason="", readiness=false. Elapsed: 3.777418ms - Dec 14 15:44:48.026: INFO: Pod "pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014151195s - Dec 14 15:44:50.024: INFO: Pod "pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012026087s - STEP: Saw pod success 12/14/22 15:44:50.024 - Dec 14 15:44:50.025: INFO: Pod "pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1" satisfied condition "Succeeded or Failed" - Dec 14 15:44:50.030: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1 container agnhost-container: - STEP: delete the pod 12/14/22 15:44:50.043 - Dec 14 15:44:50.065: INFO: Waiting for pod pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1 to disappear - Dec 14 15:44:50.071: INFO: Pod pod-projected-configmaps-cfe7a1d2-7435-481b-9ed8-fa3825f1d1c1 no longer exists - [AfterEach] [sig-storage] Projected configMap + STEP: Creating a kubernetes client 02/25/23 09:50:45.691 + Feb 25 09:50:45.691: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename pods 02/25/23 09:50:45.694 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:50:45.717 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:50:45.72 + [BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 + [It] should get a host IP [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:203 + STEP: creating pod 02/25/23 09:50:45.724 + Feb 25 09:50:45.740: INFO: Waiting up to 5m0s for pod "pod-hostip-f26c78f5-20f9-4273-a7e5-cbbb4fc5e963" in namespace "pods-6602" to be "running and ready" + Feb 25 09:50:45.746: INFO: Pod "pod-hostip-f26c78f5-20f9-4273-a7e5-cbbb4fc5e963": Phase="Pending", Reason="", readiness=false. Elapsed: 5.62358ms + Feb 25 09:50:45.746: INFO: The phase of Pod pod-hostip-f26c78f5-20f9-4273-a7e5-cbbb4fc5e963 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 09:50:47.752: INFO: Pod "pod-hostip-f26c78f5-20f9-4273-a7e5-cbbb4fc5e963": Phase="Running", Reason="", readiness=true. Elapsed: 2.011449349s + Feb 25 09:50:47.752: INFO: The phase of Pod pod-hostip-f26c78f5-20f9-4273-a7e5-cbbb4fc5e963 is Running (Ready = true) + Feb 25 09:50:47.752: INFO: Pod "pod-hostip-f26c78f5-20f9-4273-a7e5-cbbb4fc5e963" satisfied condition "running and ready" + Feb 25 09:50:47.763: INFO: Pod pod-hostip-f26c78f5-20f9-4273-a7e5-cbbb4fc5e963 has hostIP: 192.168.121.117 + [AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 - Dec 14 15:44:50.071: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-762" for this suite. 12/14/22 15:44:50.077 + Feb 25 09:50:47.764: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "pods-6602" for this suite. 02/25/23 09:50:47.772 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSS ------------------------------ -[sig-scheduling] SchedulerPredicates [Serial] - validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] - test/e2e/scheduling/predicates.go:699 -[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] +[sig-apps] Job + should apply changes to a job status [Conformance] + test/e2e/apps/job.go:464 +[BeforeEach] [sig-apps] Job test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:44:50.103 -Dec 14 15:44:50.103: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename sched-pred 12/14/22 15:44:50.105 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:44:50.139 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:44:50.145 -[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:92 -Dec 14 15:44:50.150: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready -Dec 14 15:44:50.162: INFO: Waiting for terminating namespaces to be deleted... -Dec 14 15:44:50.168: INFO: -Logging pods the apiserver thinks is on node kiedai4iakae-1 before test -Dec 14 15:44:50.177: INFO: kube-flannel-ds-d6vqv from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.178: INFO: Container kube-flannel ready: true, restart count 0 -Dec 14 15:44:50.178: INFO: kube-addon-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.178: INFO: Container kube-addon-manager ready: true, restart count 1 -Dec 14 15:44:50.178: INFO: kube-apiserver-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.178: INFO: Container kube-apiserver ready: true, restart count 1 -Dec 14 15:44:50.178: INFO: kube-controller-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:29:20 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.179: INFO: Container kube-controller-manager ready: true, restart count 2 -Dec 14 15:44:50.179: INFO: kube-proxy-fs5fm from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.179: INFO: Container kube-proxy ready: true, restart count 0 -Dec 14 15:44:50.179: INFO: kube-scheduler-kiedai4iakae-1 from kube-system started at 2022-12-14 15:25:59 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.179: INFO: Container kube-scheduler ready: true, restart count 2 -Dec 14 15:44:50.179: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-z7zks from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 15:44:50.179: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 15:44:50.180: INFO: Container systemd-logs ready: true, restart count 0 -Dec 14 15:44:50.180: INFO: -Logging pods the apiserver thinks is on node kiedai4iakae-2 before test -Dec 14 15:44:50.189: INFO: kube-flannel-ds-nvdqm from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.189: INFO: Container kube-flannel ready: true, restart count 0 -Dec 14 15:44:50.189: INFO: coredns-565d847f94-xhv9r from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.189: INFO: Container coredns ready: true, restart count 0 -Dec 14 15:44:50.189: INFO: kube-addon-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.189: INFO: Container kube-addon-manager ready: true, restart count 1 -Dec 14 15:44:50.189: INFO: kube-apiserver-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.189: INFO: Container kube-apiserver ready: true, restart count 1 -Dec 14 15:44:50.189: INFO: kube-controller-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.189: INFO: Container kube-controller-manager ready: true, restart count 1 -Dec 14 15:44:50.189: INFO: kube-proxy-dqjw5 from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.189: INFO: Container kube-proxy ready: true, restart count 0 -Dec 14 15:44:50.189: INFO: kube-scheduler-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.189: INFO: Container kube-scheduler ready: true, restart count 2 -Dec 14 15:44:50.189: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-cgz78 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 15:44:50.189: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 15:44:50.189: INFO: Container systemd-logs ready: true, restart count 0 -Dec 14 15:44:50.189: INFO: -Logging pods the apiserver thinks is on node kiedai4iakae-3 before test -Dec 14 15:44:50.200: INFO: kube-flannel-ds-j6hx4 from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.200: INFO: Container kube-flannel ready: true, restart count 0 -Dec 14 15:44:50.200: INFO: coredns-565d847f94-6tl8c from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.200: INFO: Container coredns ready: true, restart count 0 -Dec 14 15:44:50.200: INFO: kube-proxy-8kpmb from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.200: INFO: Container kube-proxy ready: true, restart count 0 -Dec 14 15:44:50.200: INFO: sonobuoy from sonobuoy started at 2022-12-14 15:35:59 +0000 UTC (1 container statuses recorded) -Dec 14 15:44:50.200: INFO: Container kube-sonobuoy ready: true, restart count 0 -Dec 14 15:44:50.200: INFO: sonobuoy-e2e-job-177ae5e9cdd34bac from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 15:44:50.200: INFO: Container e2e ready: true, restart count 0 -Dec 14 15:44:50.200: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 15:44:50.200: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-2fkm5 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 15:44:50.200: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 15:44:50.200: INFO: Container systemd-logs ready: true, restart count 0 -[It] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] - test/e2e/scheduling/predicates.go:699 -STEP: Trying to launch a pod without a label to get a node which can launch it. 12/14/22 15:44:50.2 -Dec 14 15:44:50.215: INFO: Waiting up to 1m0s for pod "without-label" in namespace "sched-pred-5652" to be "running" -Dec 14 15:44:50.220: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 4.351787ms -Dec 14 15:44:52.228: INFO: Pod "without-label": Phase="Running", Reason="", readiness=true. Elapsed: 2.012466064s -Dec 14 15:44:52.228: INFO: Pod "without-label" satisfied condition "running" -STEP: Explicitly delete pod here to free the resource it takes. 12/14/22 15:44:52.233 -STEP: Trying to apply a random label on the found node. 12/14/22 15:44:52.257 -STEP: verifying the node has the label kubernetes.io/e2e-c1531eaa-e48b-4250-ac5c-5bbab4acfce3 95 12/14/22 15:44:52.272 -STEP: Trying to create a pod(pod4) with hostport 54322 and hostIP 0.0.0.0(empty string here) and expect scheduled 12/14/22 15:44:52.282 -Dec 14 15:44:52.291: INFO: Waiting up to 5m0s for pod "pod4" in namespace "sched-pred-5652" to be "not pending" -Dec 14 15:44:52.299: INFO: Pod "pod4": Phase="Pending", Reason="", readiness=false. Elapsed: 7.536083ms -Dec 14 15:44:54.309: INFO: Pod "pod4": Phase="Running", Reason="", readiness=true. Elapsed: 2.017820618s -Dec 14 15:44:54.309: INFO: Pod "pod4" satisfied condition "not pending" -STEP: Trying to create another pod(pod5) with hostport 54322 but hostIP 192.168.121.244 on the node which pod4 resides and expect not scheduled 12/14/22 15:44:54.309 -Dec 14 15:44:54.318: INFO: Waiting up to 5m0s for pod "pod5" in namespace "sched-pred-5652" to be "not pending" -Dec 14 15:44:54.322: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3.836692ms -Dec 14 15:44:56.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013914492s -Dec 14 15:44:58.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4.013662615s -Dec 14 15:45:00.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 6.012612921s -Dec 14 15:45:02.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 8.015287733s -Dec 14 15:45:04.336: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 10.017813444s -Dec 14 15:45:06.328: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 12.009360933s -Dec 14 15:45:08.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 14.012024986s -Dec 14 15:45:10.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 16.013587625s -Dec 14 15:45:12.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 18.01155196s -Dec 14 15:45:14.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 20.012111409s -Dec 14 15:45:16.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 22.015107698s -Dec 14 15:45:18.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 24.012895866s -Dec 14 15:45:20.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 26.013362186s -Dec 14 15:45:22.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 28.014679248s -Dec 14 15:45:24.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 30.011477799s -Dec 14 15:45:26.335: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 32.016541437s -Dec 14 15:45:28.335: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 34.016136131s -Dec 14 15:45:30.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 36.012765782s -Dec 14 15:45:32.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 38.013682284s -Dec 14 15:45:34.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 40.012884062s -Dec 14 15:45:36.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 42.011620303s -Dec 14 15:45:38.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 44.012409161s -Dec 14 15:45:40.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 46.012775679s -Dec 14 15:45:42.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 48.012819316s -Dec 14 15:45:44.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 50.014740669s -Dec 14 15:45:46.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 52.012760766s -Dec 14 15:45:48.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 54.013058078s -Dec 14 15:45:50.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 56.015406244s -Dec 14 15:45:52.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 58.011620362s -Dec 14 15:45:54.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m0.012183518s -Dec 14 15:45:56.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m2.012859143s -Dec 14 15:45:58.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m4.016128136s -Dec 14 15:46:00.328: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m6.009551558s -Dec 14 15:46:02.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m8.013465291s -Dec 14 15:46:04.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m10.011290602s -Dec 14 15:46:06.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m12.012805141s -Dec 14 15:46:08.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m14.012797662s -Dec 14 15:46:10.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m16.010920071s -Dec 14 15:46:12.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m18.011224855s -Dec 14 15:46:14.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m20.013648854s -Dec 14 15:46:16.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m22.014302598s -Dec 14 15:46:18.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m24.011636798s -Dec 14 15:46:20.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m26.013678817s -Dec 14 15:46:22.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m28.012860483s -Dec 14 15:46:24.339: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m30.020331067s -Dec 14 15:46:26.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m32.013681589s -Dec 14 15:46:28.328: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m34.010120669s -Dec 14 15:46:30.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m36.01476053s -Dec 14 15:46:32.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m38.014167367s -Dec 14 15:46:34.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m40.012149406s -Dec 14 15:46:36.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m42.01464694s -Dec 14 15:46:38.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m44.01458974s -Dec 14 15:46:40.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m46.010473841s -Dec 14 15:46:42.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m48.011451812s -Dec 14 15:46:44.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m50.01337125s -Dec 14 15:46:46.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m52.010823958s -Dec 14 15:46:48.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m54.013547405s -Dec 14 15:46:50.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m56.013298901s -Dec 14 15:46:52.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m58.01460729s -Dec 14 15:46:54.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m0.015615087s -Dec 14 15:46:56.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m2.014435863s -Dec 14 15:46:58.335: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m4.016521327s -Dec 14 15:47:00.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m6.012626845s -Dec 14 15:47:02.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m8.010619764s -Dec 14 15:47:04.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m10.01098207s -Dec 14 15:47:06.337: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m12.018386991s -Dec 14 15:47:08.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m14.010175185s -Dec 14 15:47:10.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m16.011636995s -Dec 14 15:47:12.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m18.01516145s -Dec 14 15:47:14.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m20.012064419s -Dec 14 15:47:16.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m22.013764947s -Dec 14 15:47:18.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m24.014771978s -Dec 14 15:47:20.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m26.013884307s -Dec 14 15:47:22.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m28.012113719s -Dec 14 15:47:24.337: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m30.018319309s -Dec 14 15:47:26.328: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m32.009773183s -Dec 14 15:47:28.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m34.012381933s -Dec 14 15:47:30.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m36.0131954s -Dec 14 15:47:32.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m38.012840872s -Dec 14 15:47:34.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m40.014799997s -Dec 14 15:47:36.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m42.013385637s -Dec 14 15:47:38.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m44.012277012s -Dec 14 15:47:40.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m46.010175359s -Dec 14 15:47:42.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m48.013512672s -Dec 14 15:47:44.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m50.01022529s -Dec 14 15:47:46.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m52.013428271s -Dec 14 15:47:48.338: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m54.019599854s -Dec 14 15:47:50.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m56.011779726s -Dec 14 15:47:52.360: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m58.041993431s -Dec 14 15:47:54.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m0.014610969s -Dec 14 15:47:56.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m2.011619074s -Dec 14 15:47:58.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m4.013212575s -Dec 14 15:48:00.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m6.013624313s -Dec 14 15:48:02.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m8.010685946s -Dec 14 15:48:04.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m10.012127896s -Dec 14 15:48:06.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m12.015201853s -Dec 14 15:48:08.327: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m14.009126992s -Dec 14 15:48:10.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m16.01377171s -Dec 14 15:48:12.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m18.014243484s -Dec 14 15:48:14.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m20.01316823s -Dec 14 15:48:16.341: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m22.02282544s -Dec 14 15:48:18.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m24.014353519s -Dec 14 15:48:20.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m26.011320896s -Dec 14 15:48:22.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m28.014803354s -Dec 14 15:48:24.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m30.014712844s -Dec 14 15:48:26.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m32.011618734s -Dec 14 15:48:28.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m34.016078537s -Dec 14 15:48:30.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m36.0135127s -Dec 14 15:48:32.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m38.010358304s -Dec 14 15:48:34.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m40.013443933s -Dec 14 15:48:36.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m42.013768886s -Dec 14 15:48:38.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m44.015663324s -Dec 14 15:48:40.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m46.01520507s -Dec 14 15:48:42.345: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m48.026683235s -Dec 14 15:48:44.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m50.013113275s -Dec 14 15:48:46.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m52.013107119s -Dec 14 15:48:48.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m54.014294367s -Dec 14 15:48:50.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m56.010999787s -Dec 14 15:48:52.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m58.012939155s -Dec 14 15:48:54.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m0.012399426s -Dec 14 15:48:56.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m2.013629376s -Dec 14 15:48:58.335: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m4.016236696s -Dec 14 15:49:00.328: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m6.009782982s -Dec 14 15:49:02.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m8.01307944s -Dec 14 15:49:04.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m10.011167926s -Dec 14 15:49:06.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m12.012565123s -Dec 14 15:49:08.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m14.013067447s -Dec 14 15:49:10.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m16.012349886s -Dec 14 15:49:12.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m18.013253623s -Dec 14 15:49:14.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m20.011375699s -Dec 14 15:49:16.336: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m22.017201106s -Dec 14 15:49:18.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m24.015875869s -Dec 14 15:49:20.349: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m26.031115165s -Dec 14 15:49:22.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m28.014433601s -Dec 14 15:49:24.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m30.012904265s -Dec 14 15:49:26.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m32.011837335s -Dec 14 15:49:28.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m34.014922683s -Dec 14 15:49:30.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m36.016065277s -Dec 14 15:49:32.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m38.013073156s -Dec 14 15:49:34.335: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m40.016988662s -Dec 14 15:49:36.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m42.014715217s -Dec 14 15:49:38.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m44.012406715s -Dec 14 15:49:40.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m46.014678299s -Dec 14 15:49:42.328: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m48.009385978s -Dec 14 15:49:44.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m50.012958423s -Dec 14 15:49:46.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m52.012743354s -Dec 14 15:49:48.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m54.014152341s -Dec 14 15:49:50.343: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m56.024320403s -Dec 14 15:49:52.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m58.01239293s -Dec 14 15:49:54.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 5m0.013072196s -Dec 14 15:49:54.338: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 5m0.019641258s -STEP: removing the label kubernetes.io/e2e-c1531eaa-e48b-4250-ac5c-5bbab4acfce3 off the node kiedai4iakae-3 12/14/22 15:49:54.338 -STEP: verifying the node doesn't have the label kubernetes.io/e2e-c1531eaa-e48b-4250-ac5c-5bbab4acfce3 12/14/22 15:49:54.36 -[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] +STEP: Creating a kubernetes client 02/25/23 09:50:47.794 +Feb 25 09:50:47.796: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename job 02/25/23 09:50:47.8 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:50:47.829 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:50:47.835 +[It] should apply changes to a job status [Conformance] + test/e2e/apps/job.go:464 +STEP: Creating a job 02/25/23 09:50:47.84 +STEP: Ensure pods equal to paralellism count is attached to the job 02/25/23 09:50:47.85 +STEP: patching /status 02/25/23 09:50:51.859 +STEP: updating /status 02/25/23 09:50:51.876 +STEP: get /status 02/25/23 09:50:51.923 +[AfterEach] [sig-apps] Job test/e2e/framework/framework.go:187 -Dec 14 15:49:54.369: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "sched-pred-5652" for this suite. 12/14/22 15:49:54.378 -[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:83 -{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]","completed":19,"skipped":303,"failed":0} +Feb 25 09:50:51.930: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "job-6582" for this suite. 02/25/23 09:50:51.945 +{"msg":"PASSED [sig-apps] Job should apply changes to a job status [Conformance]","completed":19,"skipped":530,"failed":0} ------------------------------ -• [SLOW TEST] [304.287 seconds] -[sig-scheduling] SchedulerPredicates [Serial] -test/e2e/scheduling/framework.go:40 - validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] - test/e2e/scheduling/predicates.go:699 +• [4.162 seconds] +[sig-apps] Job +test/e2e/apps/framework.go:23 + should apply changes to a job status [Conformance] + test/e2e/apps/job.go:464 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + [BeforeEach] [sig-apps] Job test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:44:50.103 - Dec 14 15:44:50.103: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename sched-pred 12/14/22 15:44:50.105 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:44:50.139 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:44:50.145 - [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:92 - Dec 14 15:44:50.150: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready - Dec 14 15:44:50.162: INFO: Waiting for terminating namespaces to be deleted... - Dec 14 15:44:50.168: INFO: - Logging pods the apiserver thinks is on node kiedai4iakae-1 before test - Dec 14 15:44:50.177: INFO: kube-flannel-ds-d6vqv from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.178: INFO: Container kube-flannel ready: true, restart count 0 - Dec 14 15:44:50.178: INFO: kube-addon-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.178: INFO: Container kube-addon-manager ready: true, restart count 1 - Dec 14 15:44:50.178: INFO: kube-apiserver-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.178: INFO: Container kube-apiserver ready: true, restart count 1 - Dec 14 15:44:50.178: INFO: kube-controller-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:29:20 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.179: INFO: Container kube-controller-manager ready: true, restart count 2 - Dec 14 15:44:50.179: INFO: kube-proxy-fs5fm from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.179: INFO: Container kube-proxy ready: true, restart count 0 - Dec 14 15:44:50.179: INFO: kube-scheduler-kiedai4iakae-1 from kube-system started at 2022-12-14 15:25:59 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.179: INFO: Container kube-scheduler ready: true, restart count 2 - Dec 14 15:44:50.179: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-z7zks from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 15:44:50.179: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 15:44:50.180: INFO: Container systemd-logs ready: true, restart count 0 - Dec 14 15:44:50.180: INFO: - Logging pods the apiserver thinks is on node kiedai4iakae-2 before test - Dec 14 15:44:50.189: INFO: kube-flannel-ds-nvdqm from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.189: INFO: Container kube-flannel ready: true, restart count 0 - Dec 14 15:44:50.189: INFO: coredns-565d847f94-xhv9r from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.189: INFO: Container coredns ready: true, restart count 0 - Dec 14 15:44:50.189: INFO: kube-addon-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.189: INFO: Container kube-addon-manager ready: true, restart count 1 - Dec 14 15:44:50.189: INFO: kube-apiserver-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.189: INFO: Container kube-apiserver ready: true, restart count 1 - Dec 14 15:44:50.189: INFO: kube-controller-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.189: INFO: Container kube-controller-manager ready: true, restart count 1 - Dec 14 15:44:50.189: INFO: kube-proxy-dqjw5 from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.189: INFO: Container kube-proxy ready: true, restart count 0 - Dec 14 15:44:50.189: INFO: kube-scheduler-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.189: INFO: Container kube-scheduler ready: true, restart count 2 - Dec 14 15:44:50.189: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-cgz78 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 15:44:50.189: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 15:44:50.189: INFO: Container systemd-logs ready: true, restart count 0 - Dec 14 15:44:50.189: INFO: - Logging pods the apiserver thinks is on node kiedai4iakae-3 before test - Dec 14 15:44:50.200: INFO: kube-flannel-ds-j6hx4 from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.200: INFO: Container kube-flannel ready: true, restart count 0 - Dec 14 15:44:50.200: INFO: coredns-565d847f94-6tl8c from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.200: INFO: Container coredns ready: true, restart count 0 - Dec 14 15:44:50.200: INFO: kube-proxy-8kpmb from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.200: INFO: Container kube-proxy ready: true, restart count 0 - Dec 14 15:44:50.200: INFO: sonobuoy from sonobuoy started at 2022-12-14 15:35:59 +0000 UTC (1 container statuses recorded) - Dec 14 15:44:50.200: INFO: Container kube-sonobuoy ready: true, restart count 0 - Dec 14 15:44:50.200: INFO: sonobuoy-e2e-job-177ae5e9cdd34bac from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 15:44:50.200: INFO: Container e2e ready: true, restart count 0 - Dec 14 15:44:50.200: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 15:44:50.200: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-2fkm5 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 15:44:50.200: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 15:44:50.200: INFO: Container systemd-logs ready: true, restart count 0 - [It] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] - test/e2e/scheduling/predicates.go:699 - STEP: Trying to launch a pod without a label to get a node which can launch it. 12/14/22 15:44:50.2 - Dec 14 15:44:50.215: INFO: Waiting up to 1m0s for pod "without-label" in namespace "sched-pred-5652" to be "running" - Dec 14 15:44:50.220: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 4.351787ms - Dec 14 15:44:52.228: INFO: Pod "without-label": Phase="Running", Reason="", readiness=true. Elapsed: 2.012466064s - Dec 14 15:44:52.228: INFO: Pod "without-label" satisfied condition "running" - STEP: Explicitly delete pod here to free the resource it takes. 12/14/22 15:44:52.233 - STEP: Trying to apply a random label on the found node. 12/14/22 15:44:52.257 - STEP: verifying the node has the label kubernetes.io/e2e-c1531eaa-e48b-4250-ac5c-5bbab4acfce3 95 12/14/22 15:44:52.272 - STEP: Trying to create a pod(pod4) with hostport 54322 and hostIP 0.0.0.0(empty string here) and expect scheduled 12/14/22 15:44:52.282 - Dec 14 15:44:52.291: INFO: Waiting up to 5m0s for pod "pod4" in namespace "sched-pred-5652" to be "not pending" - Dec 14 15:44:52.299: INFO: Pod "pod4": Phase="Pending", Reason="", readiness=false. Elapsed: 7.536083ms - Dec 14 15:44:54.309: INFO: Pod "pod4": Phase="Running", Reason="", readiness=true. Elapsed: 2.017820618s - Dec 14 15:44:54.309: INFO: Pod "pod4" satisfied condition "not pending" - STEP: Trying to create another pod(pod5) with hostport 54322 but hostIP 192.168.121.244 on the node which pod4 resides and expect not scheduled 12/14/22 15:44:54.309 - Dec 14 15:44:54.318: INFO: Waiting up to 5m0s for pod "pod5" in namespace "sched-pred-5652" to be "not pending" - Dec 14 15:44:54.322: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3.836692ms - Dec 14 15:44:56.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013914492s - Dec 14 15:44:58.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4.013662615s - Dec 14 15:45:00.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 6.012612921s - Dec 14 15:45:02.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 8.015287733s - Dec 14 15:45:04.336: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 10.017813444s - Dec 14 15:45:06.328: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 12.009360933s - Dec 14 15:45:08.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 14.012024986s - Dec 14 15:45:10.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 16.013587625s - Dec 14 15:45:12.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 18.01155196s - Dec 14 15:45:14.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 20.012111409s - Dec 14 15:45:16.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 22.015107698s - Dec 14 15:45:18.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 24.012895866s - Dec 14 15:45:20.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 26.013362186s - Dec 14 15:45:22.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 28.014679248s - Dec 14 15:45:24.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 30.011477799s - Dec 14 15:45:26.335: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 32.016541437s - Dec 14 15:45:28.335: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 34.016136131s - Dec 14 15:45:30.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 36.012765782s - Dec 14 15:45:32.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 38.013682284s - Dec 14 15:45:34.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 40.012884062s - Dec 14 15:45:36.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 42.011620303s - Dec 14 15:45:38.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 44.012409161s - Dec 14 15:45:40.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 46.012775679s - Dec 14 15:45:42.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 48.012819316s - Dec 14 15:45:44.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 50.014740669s - Dec 14 15:45:46.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 52.012760766s - Dec 14 15:45:48.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 54.013058078s - Dec 14 15:45:50.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 56.015406244s - Dec 14 15:45:52.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 58.011620362s - Dec 14 15:45:54.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m0.012183518s - Dec 14 15:45:56.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m2.012859143s - Dec 14 15:45:58.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m4.016128136s - Dec 14 15:46:00.328: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m6.009551558s - Dec 14 15:46:02.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m8.013465291s - Dec 14 15:46:04.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m10.011290602s - Dec 14 15:46:06.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m12.012805141s - Dec 14 15:46:08.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m14.012797662s - Dec 14 15:46:10.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m16.010920071s - Dec 14 15:46:12.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m18.011224855s - Dec 14 15:46:14.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m20.013648854s - Dec 14 15:46:16.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m22.014302598s - Dec 14 15:46:18.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m24.011636798s - Dec 14 15:46:20.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m26.013678817s - Dec 14 15:46:22.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m28.012860483s - Dec 14 15:46:24.339: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m30.020331067s - Dec 14 15:46:26.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m32.013681589s - Dec 14 15:46:28.328: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m34.010120669s - Dec 14 15:46:30.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m36.01476053s - Dec 14 15:46:32.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m38.014167367s - Dec 14 15:46:34.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m40.012149406s - Dec 14 15:46:36.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m42.01464694s - Dec 14 15:46:38.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m44.01458974s - Dec 14 15:46:40.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m46.010473841s - Dec 14 15:46:42.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m48.011451812s - Dec 14 15:46:44.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m50.01337125s - Dec 14 15:46:46.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m52.010823958s - Dec 14 15:46:48.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m54.013547405s - Dec 14 15:46:50.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m56.013298901s - Dec 14 15:46:52.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m58.01460729s - Dec 14 15:46:54.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m0.015615087s - Dec 14 15:46:56.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m2.014435863s - Dec 14 15:46:58.335: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m4.016521327s - Dec 14 15:47:00.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m6.012626845s - Dec 14 15:47:02.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m8.010619764s - Dec 14 15:47:04.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m10.01098207s - Dec 14 15:47:06.337: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m12.018386991s - Dec 14 15:47:08.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m14.010175185s - Dec 14 15:47:10.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m16.011636995s - Dec 14 15:47:12.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m18.01516145s - Dec 14 15:47:14.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m20.012064419s - Dec 14 15:47:16.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m22.013764947s - Dec 14 15:47:18.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m24.014771978s - Dec 14 15:47:20.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m26.013884307s - Dec 14 15:47:22.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m28.012113719s - Dec 14 15:47:24.337: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m30.018319309s - Dec 14 15:47:26.328: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m32.009773183s - Dec 14 15:47:28.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m34.012381933s - Dec 14 15:47:30.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m36.0131954s - Dec 14 15:47:32.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m38.012840872s - Dec 14 15:47:34.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m40.014799997s - Dec 14 15:47:36.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m42.013385637s - Dec 14 15:47:38.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m44.012277012s - Dec 14 15:47:40.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m46.010175359s - Dec 14 15:47:42.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m48.013512672s - Dec 14 15:47:44.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m50.01022529s - Dec 14 15:47:46.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m52.013428271s - Dec 14 15:47:48.338: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m54.019599854s - Dec 14 15:47:50.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m56.011779726s - Dec 14 15:47:52.360: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m58.041993431s - Dec 14 15:47:54.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m0.014610969s - Dec 14 15:47:56.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m2.011619074s - Dec 14 15:47:58.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m4.013212575s - Dec 14 15:48:00.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m6.013624313s - Dec 14 15:48:02.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m8.010685946s - Dec 14 15:48:04.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m10.012127896s - Dec 14 15:48:06.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m12.015201853s - Dec 14 15:48:08.327: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m14.009126992s - Dec 14 15:48:10.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m16.01377171s - Dec 14 15:48:12.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m18.014243484s - Dec 14 15:48:14.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m20.01316823s - Dec 14 15:48:16.341: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m22.02282544s - Dec 14 15:48:18.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m24.014353519s - Dec 14 15:48:20.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m26.011320896s - Dec 14 15:48:22.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m28.014803354s - Dec 14 15:48:24.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m30.014712844s - Dec 14 15:48:26.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m32.011618734s - Dec 14 15:48:28.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m34.016078537s - Dec 14 15:48:30.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m36.0135127s - Dec 14 15:48:32.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m38.010358304s - Dec 14 15:48:34.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m40.013443933s - Dec 14 15:48:36.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m42.013768886s - Dec 14 15:48:38.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m44.015663324s - Dec 14 15:48:40.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m46.01520507s - Dec 14 15:48:42.345: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m48.026683235s - Dec 14 15:48:44.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m50.013113275s - Dec 14 15:48:46.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m52.013107119s - Dec 14 15:48:48.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m54.014294367s - Dec 14 15:48:50.329: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m56.010999787s - Dec 14 15:48:52.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m58.012939155s - Dec 14 15:48:54.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m0.012399426s - Dec 14 15:48:56.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m2.013629376s - Dec 14 15:48:58.335: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m4.016236696s - Dec 14 15:49:00.328: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m6.009782982s - Dec 14 15:49:02.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m8.01307944s - Dec 14 15:49:04.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m10.011167926s - Dec 14 15:49:06.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m12.012565123s - Dec 14 15:49:08.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m14.013067447s - Dec 14 15:49:10.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m16.012349886s - Dec 14 15:49:12.332: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m18.013253623s - Dec 14 15:49:14.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m20.011375699s - Dec 14 15:49:16.336: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m22.017201106s - Dec 14 15:49:18.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m24.015875869s - Dec 14 15:49:20.349: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m26.031115165s - Dec 14 15:49:22.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m28.014433601s - Dec 14 15:49:24.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m30.012904265s - Dec 14 15:49:26.330: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m32.011837335s - Dec 14 15:49:28.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m34.014922683s - Dec 14 15:49:30.334: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m36.016065277s - Dec 14 15:49:32.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m38.013073156s - Dec 14 15:49:34.335: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m40.016988662s - Dec 14 15:49:36.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m42.014715217s - Dec 14 15:49:38.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m44.012406715s - Dec 14 15:49:40.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m46.014678299s - Dec 14 15:49:42.328: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m48.009385978s - Dec 14 15:49:44.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m50.012958423s - Dec 14 15:49:46.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m52.012743354s - Dec 14 15:49:48.333: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m54.014152341s - Dec 14 15:49:50.343: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m56.024320403s - Dec 14 15:49:52.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m58.01239293s - Dec 14 15:49:54.331: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 5m0.013072196s - Dec 14 15:49:54.338: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 5m0.019641258s - STEP: removing the label kubernetes.io/e2e-c1531eaa-e48b-4250-ac5c-5bbab4acfce3 off the node kiedai4iakae-3 12/14/22 15:49:54.338 - STEP: verifying the node doesn't have the label kubernetes.io/e2e-c1531eaa-e48b-4250-ac5c-5bbab4acfce3 12/14/22 15:49:54.36 - [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + STEP: Creating a kubernetes client 02/25/23 09:50:47.794 + Feb 25 09:50:47.796: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename job 02/25/23 09:50:47.8 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:50:47.829 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:50:47.835 + [It] should apply changes to a job status [Conformance] + test/e2e/apps/job.go:464 + STEP: Creating a job 02/25/23 09:50:47.84 + STEP: Ensure pods equal to paralellism count is attached to the job 02/25/23 09:50:47.85 + STEP: patching /status 02/25/23 09:50:51.859 + STEP: updating /status 02/25/23 09:50:51.876 + STEP: get /status 02/25/23 09:50:51.923 + [AfterEach] [sig-apps] Job test/e2e/framework/framework.go:187 - Dec 14 15:49:54.369: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "sched-pred-5652" for this suite. 12/14/22 15:49:54.378 - [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:83 + Feb 25 09:50:51.930: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "job-6582" for this suite. 02/25/23 09:50:51.945 << End Captured GinkgoWriter Output ------------------------------ -SSSSS +SSSSSSSS ------------------------------ -[sig-node] Security Context when creating containers with AllowPrivilegeEscalation - should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:608 -[BeforeEach] [sig-node] Security Context +[sig-auth] Certificates API [Privileged:ClusterAdmin] + should support CSR API operations [Conformance] + test/e2e/auth/certificates.go:200 +[BeforeEach] [sig-auth] Certificates API [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:49:54.391 -Dec 14 15:49:54.391: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename security-context-test 12/14/22 15:49:54.401 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:49:54.429 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:49:54.432 -[BeforeEach] [sig-node] Security Context - test/e2e/common/node/security_context.go:49 -[It] should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:608 -Dec 14 15:49:54.448: INFO: Waiting up to 5m0s for pod "alpine-nnp-false-a253dcab-ecf9-46c8-842d-611429d5058c" in namespace "security-context-test-8454" to be "Succeeded or Failed" -Dec 14 15:49:54.453: INFO: Pod "alpine-nnp-false-a253dcab-ecf9-46c8-842d-611429d5058c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.101936ms -Dec 14 15:49:56.459: INFO: Pod "alpine-nnp-false-a253dcab-ecf9-46c8-842d-611429d5058c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010052056s -Dec 14 15:49:58.463: INFO: Pod "alpine-nnp-false-a253dcab-ecf9-46c8-842d-611429d5058c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.014319436s -Dec 14 15:50:00.460: INFO: Pod "alpine-nnp-false-a253dcab-ecf9-46c8-842d-611429d5058c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.011312687s -Dec 14 15:50:00.460: INFO: Pod "alpine-nnp-false-a253dcab-ecf9-46c8-842d-611429d5058c" satisfied condition "Succeeded or Failed" -[AfterEach] [sig-node] Security Context +STEP: Creating a kubernetes client 02/25/23 09:50:51.969 +Feb 25 09:50:51.970: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename certificates 02/25/23 09:50:51.973 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:50:51.999 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:50:52.007 +[It] should support CSR API operations [Conformance] + test/e2e/auth/certificates.go:200 +STEP: getting /apis 02/25/23 09:50:52.882 +STEP: getting /apis/certificates.k8s.io 02/25/23 09:50:52.885 +STEP: getting /apis/certificates.k8s.io/v1 02/25/23 09:50:52.886 +STEP: creating 02/25/23 09:50:52.888 +STEP: getting 02/25/23 09:50:52.921 +STEP: listing 02/25/23 09:50:52.932 +STEP: watching 02/25/23 09:50:52.939 +Feb 25 09:50:52.940: INFO: starting watch +STEP: patching 02/25/23 09:50:52.942 +STEP: updating 02/25/23 09:50:52.956 +Feb 25 09:50:52.966: INFO: waiting for watch events with expected annotations +Feb 25 09:50:52.966: INFO: saw patched and updated annotations +STEP: getting /approval 02/25/23 09:50:52.966 +STEP: patching /approval 02/25/23 09:50:52.973 +STEP: updating /approval 02/25/23 09:50:52.982 +STEP: getting /status 02/25/23 09:50:52.997 +STEP: patching /status 02/25/23 09:50:53.002 +STEP: updating /status 02/25/23 09:50:53.013 +STEP: deleting 02/25/23 09:50:53.033 +STEP: deleting a collection 02/25/23 09:50:53.065 +[AfterEach] [sig-auth] Certificates API [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 15:50:00.491: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "security-context-test-8454" for this suite. 12/14/22 15:50:00.503 -{"msg":"PASSED [sig-node] Security Context when creating containers with AllowPrivilegeEscalation should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]","completed":20,"skipped":308,"failed":0} +Feb 25 09:50:53.098: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "certificates-9187" for this suite. 02/25/23 09:50:53.104 +{"msg":"PASSED [sig-auth] Certificates API [Privileged:ClusterAdmin] should support CSR API operations [Conformance]","completed":20,"skipped":538,"failed":0} ------------------------------ -• [SLOW TEST] [6.127 seconds] -[sig-node] Security Context -test/e2e/common/node/framework.go:23 - when creating containers with AllowPrivilegeEscalation - test/e2e/common/node/security_context.go:554 - should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:608 - - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Security Context - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:49:54.391 - Dec 14 15:49:54.391: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename security-context-test 12/14/22 15:49:54.401 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:49:54.429 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:49:54.432 - [BeforeEach] [sig-node] Security Context - test/e2e/common/node/security_context.go:49 - [It] should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:608 - Dec 14 15:49:54.448: INFO: Waiting up to 5m0s for pod "alpine-nnp-false-a253dcab-ecf9-46c8-842d-611429d5058c" in namespace "security-context-test-8454" to be "Succeeded or Failed" - Dec 14 15:49:54.453: INFO: Pod "alpine-nnp-false-a253dcab-ecf9-46c8-842d-611429d5058c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.101936ms - Dec 14 15:49:56.459: INFO: Pod "alpine-nnp-false-a253dcab-ecf9-46c8-842d-611429d5058c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010052056s - Dec 14 15:49:58.463: INFO: Pod "alpine-nnp-false-a253dcab-ecf9-46c8-842d-611429d5058c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.014319436s - Dec 14 15:50:00.460: INFO: Pod "alpine-nnp-false-a253dcab-ecf9-46c8-842d-611429d5058c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.011312687s - Dec 14 15:50:00.460: INFO: Pod "alpine-nnp-false-a253dcab-ecf9-46c8-842d-611429d5058c" satisfied condition "Succeeded or Failed" - [AfterEach] [sig-node] Security Context - test/e2e/framework/framework.go:187 - Dec 14 15:50:00.491: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "security-context-test-8454" for this suite. 12/14/22 15:50:00.503 - << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-storage] Projected combined - should project all components that make up the projection API [Projection][NodeConformance] [Conformance] - test/e2e/common/storage/projected_combined.go:43 -[BeforeEach] [sig-storage] Projected combined - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:50:00.527 -Dec 14 15:50:00.528: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 15:50:00.53 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:50:00.554 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:50:00.558 -[It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] - test/e2e/common/storage/projected_combined.go:43 -STEP: Creating configMap with name configmap-projected-all-test-volume-3f36aac8-ea98-417d-af86-4fe528d44527 12/14/22 15:50:00.562 -STEP: Creating secret with name secret-projected-all-test-volume-ef6c2724-701a-47b3-a62d-185f6950371d 12/14/22 15:50:00.571 -STEP: Creating a pod to test Check all projections for projected volume plugin 12/14/22 15:50:00.576 -Dec 14 15:50:00.598: INFO: Waiting up to 5m0s for pod "projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837" in namespace "projected-4314" to be "Succeeded or Failed" -Dec 14 15:50:00.603: INFO: Pod "projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837": Phase="Pending", Reason="", readiness=false. Elapsed: 4.792711ms -Dec 14 15:50:02.614: INFO: Pod "projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016338487s -Dec 14 15:50:04.615: INFO: Pod "projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016628413s -STEP: Saw pod success 12/14/22 15:50:04.615 -Dec 14 15:50:04.615: INFO: Pod "projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837" satisfied condition "Succeeded or Failed" -Dec 14 15:50:04.621: INFO: Trying to get logs from node kiedai4iakae-3 pod projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837 container projected-all-volume-test: -STEP: delete the pod 12/14/22 15:50:04.635 -Dec 14 15:50:04.660: INFO: Waiting for pod projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837 to disappear -Dec 14 15:50:04.665: INFO: Pod projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837 no longer exists -[AfterEach] [sig-storage] Projected combined - test/e2e/framework/framework.go:187 -Dec 14 15:50:04.665: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-4314" for this suite. 12/14/22 15:50:04.672 -{"msg":"PASSED [sig-storage] Projected combined should project all components that make up the projection API [Projection][NodeConformance] [Conformance]","completed":21,"skipped":331,"failed":0} ------------------------------- -• [4.154 seconds] -[sig-storage] Projected combined -test/e2e/common/storage/framework.go:23 - should project all components that make up the projection API [Projection][NodeConformance] [Conformance] - test/e2e/common/storage/projected_combined.go:43 +• [1.145 seconds] +[sig-auth] Certificates API [Privileged:ClusterAdmin] +test/e2e/auth/framework.go:23 + should support CSR API operations [Conformance] + test/e2e/auth/certificates.go:200 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected combined + [BeforeEach] [sig-auth] Certificates API [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:50:00.527 - Dec 14 15:50:00.528: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 15:50:00.53 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:50:00.554 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:50:00.558 - [It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] - test/e2e/common/storage/projected_combined.go:43 - STEP: Creating configMap with name configmap-projected-all-test-volume-3f36aac8-ea98-417d-af86-4fe528d44527 12/14/22 15:50:00.562 - STEP: Creating secret with name secret-projected-all-test-volume-ef6c2724-701a-47b3-a62d-185f6950371d 12/14/22 15:50:00.571 - STEP: Creating a pod to test Check all projections for projected volume plugin 12/14/22 15:50:00.576 - Dec 14 15:50:00.598: INFO: Waiting up to 5m0s for pod "projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837" in namespace "projected-4314" to be "Succeeded or Failed" - Dec 14 15:50:00.603: INFO: Pod "projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837": Phase="Pending", Reason="", readiness=false. Elapsed: 4.792711ms - Dec 14 15:50:02.614: INFO: Pod "projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016338487s - Dec 14 15:50:04.615: INFO: Pod "projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016628413s - STEP: Saw pod success 12/14/22 15:50:04.615 - Dec 14 15:50:04.615: INFO: Pod "projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837" satisfied condition "Succeeded or Failed" - Dec 14 15:50:04.621: INFO: Trying to get logs from node kiedai4iakae-3 pod projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837 container projected-all-volume-test: - STEP: delete the pod 12/14/22 15:50:04.635 - Dec 14 15:50:04.660: INFO: Waiting for pod projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837 to disappear - Dec 14 15:50:04.665: INFO: Pod projected-volume-fec529d1-6ce8-4191-bd4d-704861b0d837 no longer exists - [AfterEach] [sig-storage] Projected combined + STEP: Creating a kubernetes client 02/25/23 09:50:51.969 + Feb 25 09:50:51.970: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename certificates 02/25/23 09:50:51.973 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:50:51.999 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:50:52.007 + [It] should support CSR API operations [Conformance] + test/e2e/auth/certificates.go:200 + STEP: getting /apis 02/25/23 09:50:52.882 + STEP: getting /apis/certificates.k8s.io 02/25/23 09:50:52.885 + STEP: getting /apis/certificates.k8s.io/v1 02/25/23 09:50:52.886 + STEP: creating 02/25/23 09:50:52.888 + STEP: getting 02/25/23 09:50:52.921 + STEP: listing 02/25/23 09:50:52.932 + STEP: watching 02/25/23 09:50:52.939 + Feb 25 09:50:52.940: INFO: starting watch + STEP: patching 02/25/23 09:50:52.942 + STEP: updating 02/25/23 09:50:52.956 + Feb 25 09:50:52.966: INFO: waiting for watch events with expected annotations + Feb 25 09:50:52.966: INFO: saw patched and updated annotations + STEP: getting /approval 02/25/23 09:50:52.966 + STEP: patching /approval 02/25/23 09:50:52.973 + STEP: updating /approval 02/25/23 09:50:52.982 + STEP: getting /status 02/25/23 09:50:52.997 + STEP: patching /status 02/25/23 09:50:53.002 + STEP: updating /status 02/25/23 09:50:53.013 + STEP: deleting 02/25/23 09:50:53.033 + STEP: deleting a collection 02/25/23 09:50:53.065 + [AfterEach] [sig-auth] Certificates API [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 15:50:04.665: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-4314" for this suite. 12/14/22 15:50:04.672 + Feb 25 09:50:53.098: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "certificates-9187" for this suite. 02/25/23 09:50:53.104 << End Captured GinkgoWriter Output ------------------------------ S ------------------------------ -[sig-storage] Projected secret - should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:118 -[BeforeEach] [sig-storage] Projected secret +[sig-storage] ConfigMap + updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:123 +[BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:50:04.683 -Dec 14 15:50:04.683: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 15:50:04.688 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:50:04.717 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:50:04.72 -[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:118 -STEP: Creating secret with name projected-secret-test-ea6de5bd-b1ca-49d3-87d0-2cef81a28cbe 12/14/22 15:50:04.724 -STEP: Creating a pod to test consume secrets 12/14/22 15:50:04.731 -Dec 14 15:50:04.742: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98" in namespace "projected-4677" to be "Succeeded or Failed" -Dec 14 15:50:04.747: INFO: Pod "pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98": Phase="Pending", Reason="", readiness=false. Elapsed: 4.053868ms -Dec 14 15:50:06.759: INFO: Pod "pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016544293s -Dec 14 15:50:08.756: INFO: Pod "pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013969447s -STEP: Saw pod success 12/14/22 15:50:08.757 -Dec 14 15:50:08.758: INFO: Pod "pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98" satisfied condition "Succeeded or Failed" -Dec 14 15:50:08.764: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98 container secret-volume-test: -STEP: delete the pod 12/14/22 15:50:08.775 -Dec 14 15:50:08.795: INFO: Waiting for pod pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98 to disappear -Dec 14 15:50:08.799: INFO: Pod pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98 no longer exists -[AfterEach] [sig-storage] Projected secret +STEP: Creating a kubernetes client 02/25/23 09:50:53.116 +Feb 25 09:50:53.117: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename configmap 02/25/23 09:50:53.123 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:50:53.179 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:50:53.182 +[It] updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:123 +STEP: Creating configMap with name configmap-test-upd-08163169-83d9-4936-a9c5-417af99b4f45 02/25/23 09:50:53.2 +STEP: Creating the pod 02/25/23 09:50:53.208 +Feb 25 09:50:53.221: INFO: Waiting up to 5m0s for pod "pod-configmaps-00f5ecf4-a70f-4c0d-a6f0-f75934fca279" in namespace "configmap-1502" to be "running and ready" +Feb 25 09:50:53.226: INFO: Pod "pod-configmaps-00f5ecf4-a70f-4c0d-a6f0-f75934fca279": Phase="Pending", Reason="", readiness=false. Elapsed: 5.044251ms +Feb 25 09:50:53.226: INFO: The phase of Pod pod-configmaps-00f5ecf4-a70f-4c0d-a6f0-f75934fca279 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 09:50:55.238: INFO: Pod "pod-configmaps-00f5ecf4-a70f-4c0d-a6f0-f75934fca279": Phase="Running", Reason="", readiness=true. Elapsed: 2.017075797s +Feb 25 09:50:55.238: INFO: The phase of Pod pod-configmaps-00f5ecf4-a70f-4c0d-a6f0-f75934fca279 is Running (Ready = true) +Feb 25 09:50:55.239: INFO: Pod "pod-configmaps-00f5ecf4-a70f-4c0d-a6f0-f75934fca279" satisfied condition "running and ready" +STEP: Updating configmap configmap-test-upd-08163169-83d9-4936-a9c5-417af99b4f45 02/25/23 09:50:55.255 +STEP: waiting to observe update in volume 02/25/23 09:50:55.268 +[AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 -Dec 14 15:50:08.799: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-4677" for this suite. 12/14/22 15:50:08.806 -{"msg":"PASSED [sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","completed":22,"skipped":332,"failed":0} +Feb 25 09:50:57.290: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-1502" for this suite. 02/25/23 09:50:57.307 +{"msg":"PASSED [sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance]","completed":21,"skipped":539,"failed":0} ------------------------------ -• [4.133 seconds] -[sig-storage] Projected secret +• [4.202 seconds] +[sig-storage] ConfigMap test/e2e/common/storage/framework.go:23 - should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:118 + updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:123 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected secret + [BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:50:04.683 - Dec 14 15:50:04.683: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 15:50:04.688 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:50:04.717 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:50:04.72 - [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:118 - STEP: Creating secret with name projected-secret-test-ea6de5bd-b1ca-49d3-87d0-2cef81a28cbe 12/14/22 15:50:04.724 - STEP: Creating a pod to test consume secrets 12/14/22 15:50:04.731 - Dec 14 15:50:04.742: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98" in namespace "projected-4677" to be "Succeeded or Failed" - Dec 14 15:50:04.747: INFO: Pod "pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98": Phase="Pending", Reason="", readiness=false. Elapsed: 4.053868ms - Dec 14 15:50:06.759: INFO: Pod "pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016544293s - Dec 14 15:50:08.756: INFO: Pod "pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013969447s - STEP: Saw pod success 12/14/22 15:50:08.757 - Dec 14 15:50:08.758: INFO: Pod "pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98" satisfied condition "Succeeded or Failed" - Dec 14 15:50:08.764: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98 container secret-volume-test: - STEP: delete the pod 12/14/22 15:50:08.775 - Dec 14 15:50:08.795: INFO: Waiting for pod pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98 to disappear - Dec 14 15:50:08.799: INFO: Pod pod-projected-secrets-4b7a616a-707a-499d-8ad5-8460ba064c98 no longer exists - [AfterEach] [sig-storage] Projected secret + STEP: Creating a kubernetes client 02/25/23 09:50:53.116 + Feb 25 09:50:53.117: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename configmap 02/25/23 09:50:53.123 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:50:53.179 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:50:53.182 + [It] updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:123 + STEP: Creating configMap with name configmap-test-upd-08163169-83d9-4936-a9c5-417af99b4f45 02/25/23 09:50:53.2 + STEP: Creating the pod 02/25/23 09:50:53.208 + Feb 25 09:50:53.221: INFO: Waiting up to 5m0s for pod "pod-configmaps-00f5ecf4-a70f-4c0d-a6f0-f75934fca279" in namespace "configmap-1502" to be "running and ready" + Feb 25 09:50:53.226: INFO: Pod "pod-configmaps-00f5ecf4-a70f-4c0d-a6f0-f75934fca279": Phase="Pending", Reason="", readiness=false. Elapsed: 5.044251ms + Feb 25 09:50:53.226: INFO: The phase of Pod pod-configmaps-00f5ecf4-a70f-4c0d-a6f0-f75934fca279 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 09:50:55.238: INFO: Pod "pod-configmaps-00f5ecf4-a70f-4c0d-a6f0-f75934fca279": Phase="Running", Reason="", readiness=true. Elapsed: 2.017075797s + Feb 25 09:50:55.238: INFO: The phase of Pod pod-configmaps-00f5ecf4-a70f-4c0d-a6f0-f75934fca279 is Running (Ready = true) + Feb 25 09:50:55.239: INFO: Pod "pod-configmaps-00f5ecf4-a70f-4c0d-a6f0-f75934fca279" satisfied condition "running and ready" + STEP: Updating configmap configmap-test-upd-08163169-83d9-4936-a9c5-417af99b4f45 02/25/23 09:50:55.255 + STEP: waiting to observe update in volume 02/25/23 09:50:55.268 + [AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 - Dec 14 15:50:08.799: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-4677" for this suite. 12/14/22 15:50:08.806 + Feb 25 09:50:57.290: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "configmap-1502" for this suite. 02/25/23 09:50:57.307 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSS +S ------------------------------ -[sig-storage] Downward API volume - should provide container's memory request [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:234 -[BeforeEach] [sig-storage] Downward API volume +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + should mutate configmap [Conformance] + test/e2e/apimachinery/webhook.go:251 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:50:08.822 -Dec 14 15:50:08.822: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 15:50:08.829 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:50:08.856 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:50:08.859 -[BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 -[It] should provide container's memory request [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:234 -STEP: Creating a pod to test downward API volume plugin 12/14/22 15:50:08.863 -Dec 14 15:50:08.878: INFO: Waiting up to 5m0s for pod "downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed" in namespace "downward-api-1713" to be "Succeeded or Failed" -Dec 14 15:50:08.883: INFO: Pod "downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed": Phase="Pending", Reason="", readiness=false. Elapsed: 4.487336ms -Dec 14 15:50:10.894: INFO: Pod "downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016022584s -Dec 14 15:50:12.895: INFO: Pod "downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01674486s -STEP: Saw pod success 12/14/22 15:50:12.896 -Dec 14 15:50:12.897: INFO: Pod "downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed" satisfied condition "Succeeded or Failed" -Dec 14 15:50:12.902: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed container client-container: -STEP: delete the pod 12/14/22 15:50:12.914 -Dec 14 15:50:12.937: INFO: Waiting for pod downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed to disappear -Dec 14 15:50:12.941: INFO: Pod downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed no longer exists -[AfterEach] [sig-storage] Downward API volume +STEP: Creating a kubernetes client 02/25/23 09:50:57.32 +Feb 25 09:50:57.320: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 09:50:57.322 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:50:57.351 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:50:57.356 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 +STEP: Setting up server cert 02/25/23 09:50:57.386 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 09:50:58.441 +STEP: Deploying the webhook pod 02/25/23 09:50:58.457 +STEP: Wait for the deployment to be ready 02/25/23 09:50:58.478 +Feb 25 09:50:58.496: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set +Feb 25 09:51:00.514: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 09:51:02.520: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 09:51:04.521: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 09:51:06.522: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 09:51:08.521: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} +STEP: Deploying the webhook service 02/25/23 09:51:10.532 +STEP: Verifying the service has paired with the endpoint 02/25/23 09:51:10.55 +Feb 25 09:51:11.551: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should mutate configmap [Conformance] + test/e2e/apimachinery/webhook.go:251 +STEP: Registering the mutating configmap webhook via the AdmissionRegistration API 02/25/23 09:51:11.56 +STEP: create a configmap that should be updated by the webhook 02/25/23 09:51:11.596 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 15:50:12.941: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-1713" for this suite. 12/14/22 15:50:12.947 -{"msg":"PASSED [sig-storage] Downward API volume should provide container's memory request [NodeConformance] [Conformance]","completed":23,"skipped":340,"failed":0} +Feb 25 09:51:11.642: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-6342" for this suite. 02/25/23 09:51:11.652 +STEP: Destroying namespace "webhook-6342-markers" for this suite. 02/25/23 09:51:11.666 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate configmap [Conformance]","completed":22,"skipped":540,"failed":0} ------------------------------ -• [4.134 seconds] -[sig-storage] Downward API volume -test/e2e/common/storage/framework.go:23 - should provide container's memory request [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:234 +• [SLOW TEST] [14.448 seconds] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + should mutate configmap [Conformance] + test/e2e/apimachinery/webhook.go:251 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Downward API volume + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:50:08.822 - Dec 14 15:50:08.822: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 15:50:08.829 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:50:08.856 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:50:08.859 - [BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 - [It] should provide container's memory request [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:234 - STEP: Creating a pod to test downward API volume plugin 12/14/22 15:50:08.863 - Dec 14 15:50:08.878: INFO: Waiting up to 5m0s for pod "downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed" in namespace "downward-api-1713" to be "Succeeded or Failed" - Dec 14 15:50:08.883: INFO: Pod "downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed": Phase="Pending", Reason="", readiness=false. Elapsed: 4.487336ms - Dec 14 15:50:10.894: INFO: Pod "downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016022584s - Dec 14 15:50:12.895: INFO: Pod "downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01674486s - STEP: Saw pod success 12/14/22 15:50:12.896 - Dec 14 15:50:12.897: INFO: Pod "downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed" satisfied condition "Succeeded or Failed" - Dec 14 15:50:12.902: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed container client-container: - STEP: delete the pod 12/14/22 15:50:12.914 - Dec 14 15:50:12.937: INFO: Waiting for pod downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed to disappear - Dec 14 15:50:12.941: INFO: Pod downwardapi-volume-b3d6b6bf-05c6-4986-9552-5cece5213eed no longer exists - [AfterEach] [sig-storage] Downward API volume + STEP: Creating a kubernetes client 02/25/23 09:50:57.32 + Feb 25 09:50:57.320: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 09:50:57.322 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:50:57.351 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:50:57.356 + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 + STEP: Setting up server cert 02/25/23 09:50:57.386 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 09:50:58.441 + STEP: Deploying the webhook pod 02/25/23 09:50:58.457 + STEP: Wait for the deployment to be ready 02/25/23 09:50:58.478 + Feb 25 09:50:58.496: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set + Feb 25 09:51:00.514: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 09:51:02.520: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 09:51:04.521: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 09:51:06.522: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 09:51:08.521: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 9, 50, 58, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} + STEP: Deploying the webhook service 02/25/23 09:51:10.532 + STEP: Verifying the service has paired with the endpoint 02/25/23 09:51:10.55 + Feb 25 09:51:11.551: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] should mutate configmap [Conformance] + test/e2e/apimachinery/webhook.go:251 + STEP: Registering the mutating configmap webhook via the AdmissionRegistration API 02/25/23 09:51:11.56 + STEP: create a configmap that should be updated by the webhook 02/25/23 09:51:11.596 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 15:50:12.941: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-1713" for this suite. 12/14/22 15:50:12.947 + Feb 25 09:51:11.642: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-6342" for this suite. 02/25/23 09:51:11.652 + STEP: Destroying namespace "webhook-6342-markers" for this suite. 02/25/23 09:51:11.666 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSS +SS ------------------------------ [sig-network] Services - should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance] - test/e2e/network/service.go:2189 + should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance] + test/e2e/network/service.go:2173 [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:50:12.965 -Dec 14 15:50:12.966: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 15:50:12.967 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:50:12.993 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:50:13.002 +STEP: Creating a kubernetes client 02/25/23 09:51:11.775 +Feb 25 09:51:11.776: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 09:51:11.787 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:51:11.831 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:51:11.837 [BeforeEach] [sig-network] Services test/e2e/network/service.go:758 -[It] should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance] - test/e2e/network/service.go:2189 -STEP: creating service in namespace services-7867 12/14/22 15:50:13.007 -STEP: creating service affinity-clusterip-transition in namespace services-7867 12/14/22 15:50:13.008 -STEP: creating replication controller affinity-clusterip-transition in namespace services-7867 12/14/22 15:50:13.022 -I1214 15:50:13.033614 14 runners.go:193] Created replication controller with name: affinity-clusterip-transition, namespace: services-7867, replica count: 3 -I1214 15:50:16.085581 14 runners.go:193] affinity-clusterip-transition Pods: 3 out of 3 created, 1 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -I1214 15:50:19.086726 14 runners.go:193] affinity-clusterip-transition Pods: 3 out of 3 created, 1 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -I1214 15:50:22.087940 14 runners.go:193] affinity-clusterip-transition Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -Dec 14 15:50:22.103: INFO: Creating new exec pod -Dec 14 15:50:22.119: INFO: Waiting up to 5m0s for pod "execpod-affinityqfh9z" in namespace "services-7867" to be "running" -Dec 14 15:50:22.124: INFO: Pod "execpod-affinityqfh9z": Phase="Pending", Reason="", readiness=false. Elapsed: 5.031003ms -Dec 14 15:50:24.139: INFO: Pod "execpod-affinityqfh9z": Phase="Running", Reason="", readiness=true. Elapsed: 2.020040128s -Dec 14 15:50:24.139: INFO: Pod "execpod-affinityqfh9z" satisfied condition "running" -Dec 14 15:50:25.141: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-7867 exec execpod-affinityqfh9z -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-clusterip-transition 80' -Dec 14 15:50:25.435: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-clusterip-transition 80\nConnection to affinity-clusterip-transition 80 port [tcp/http] succeeded!\n" -Dec 14 15:50:25.435: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 15:50:25.435: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-7867 exec execpod-affinityqfh9z -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.10.125 80' -Dec 14 15:50:25.671: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.10.125 80\nConnection to 10.233.10.125 80 port [tcp/http] succeeded!\n" -Dec 14 15:50:25.671: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 15:50:25.685: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-7867 exec execpod-affinityqfh9z -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.10.125:80/ ; done' -Dec 14 15:50:26.123: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n" -Dec 14 15:50:26.123: INFO: stdout: "\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-zx4lr\naffinity-clusterip-transition-h7xrx\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-zx4lr\naffinity-clusterip-transition-zx4lr\naffinity-clusterip-transition-h7xrx\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-zx4lr\naffinity-clusterip-transition-zx4lr\naffinity-clusterip-transition-h7xrx\naffinity-clusterip-transition-h7xrx\naffinity-clusterip-transition-h7xrx\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl" -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-zx4lr -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-h7xrx -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-zx4lr -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-zx4lr -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-h7xrx -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-zx4lr -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-zx4lr -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-h7xrx -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-h7xrx -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-h7xrx -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.140: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-7867 exec execpod-affinityqfh9z -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.10.125:80/ ; done' -Dec 14 15:50:26.521: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n" -Dec 14 15:50:26.521: INFO: stdout: "\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl" -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl -Dec 14 15:50:26.521: INFO: Cleaning up the exec pod -STEP: deleting ReplicationController affinity-clusterip-transition in namespace services-7867, will wait for the garbage collector to delete the pods 12/14/22 15:50:26.573 -Dec 14 15:50:26.648: INFO: Deleting ReplicationController affinity-clusterip-transition took: 13.726472ms -Dec 14 15:50:26.748: INFO: Terminating ReplicationController affinity-clusterip-transition pods took: 100.590605ms +[It] should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance] + test/e2e/network/service.go:2173 +STEP: creating service in namespace services-6402 02/25/23 09:51:11.849 +Feb 25 09:51:11.868: INFO: Waiting up to 5m0s for pod "kube-proxy-mode-detector" in namespace "services-6402" to be "running and ready" +Feb 25 09:51:11.877: INFO: Pod "kube-proxy-mode-detector": Phase="Pending", Reason="", readiness=false. Elapsed: 8.88539ms +Feb 25 09:51:11.877: INFO: The phase of Pod kube-proxy-mode-detector is Pending, waiting for it to be Running (with Ready = true) +Feb 25 09:51:13.885: INFO: Pod "kube-proxy-mode-detector": Phase="Running", Reason="", readiness=true. Elapsed: 2.016849074s +Feb 25 09:51:13.885: INFO: The phase of Pod kube-proxy-mode-detector is Running (Ready = true) +Feb 25 09:51:13.885: INFO: Pod "kube-proxy-mode-detector" satisfied condition "running and ready" +Feb 25 09:51:13.892: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6402 exec kube-proxy-mode-detector -- /bin/sh -x -c curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode' +Feb 25 09:51:14.215: INFO: stderr: "+ curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode\n" +Feb 25 09:51:14.215: INFO: stdout: "iptables" +Feb 25 09:51:14.215: INFO: proxyMode: iptables +Feb 25 09:51:14.238: INFO: Waiting for pod kube-proxy-mode-detector to disappear +Feb 25 09:51:14.244: INFO: Pod kube-proxy-mode-detector no longer exists +STEP: creating service affinity-clusterip-timeout in namespace services-6402 02/25/23 09:51:14.244 +STEP: creating replication controller affinity-clusterip-timeout in namespace services-6402 02/25/23 09:51:14.263 +I0225 09:51:14.275490 13 runners.go:193] Created replication controller with name: affinity-clusterip-timeout, namespace: services-6402, replica count: 3 +I0225 09:51:17.328761 13 runners.go:193] affinity-clusterip-timeout Pods: 3 out of 3 created, 2 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0225 09:51:20.331376 13 runners.go:193] affinity-clusterip-timeout Pods: 3 out of 3 created, 2 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0225 09:51:23.333760 13 runners.go:193] affinity-clusterip-timeout Pods: 3 out of 3 created, 2 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0225 09:51:26.335445 13 runners.go:193] affinity-clusterip-timeout Pods: 3 out of 3 created, 2 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0225 09:51:29.336009 13 runners.go:193] affinity-clusterip-timeout Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Feb 25 09:51:29.354: INFO: Creating new exec pod +Feb 25 09:51:29.365: INFO: Waiting up to 5m0s for pod "execpod-affinity45sn9" in namespace "services-6402" to be "running" +Feb 25 09:51:29.372: INFO: Pod "execpod-affinity45sn9": Phase="Pending", Reason="", readiness=false. Elapsed: 6.275666ms +Feb 25 09:51:31.382: INFO: Pod "execpod-affinity45sn9": Phase="Running", Reason="", readiness=true. Elapsed: 2.016439057s +Feb 25 09:51:31.382: INFO: Pod "execpod-affinity45sn9" satisfied condition "running" +Feb 25 09:51:32.385: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6402 exec execpod-affinity45sn9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-clusterip-timeout 80' +Feb 25 09:51:32.708: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-clusterip-timeout 80\nConnection to affinity-clusterip-timeout 80 port [tcp/http] succeeded!\n" +Feb 25 09:51:32.708: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 09:51:32.709: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6402 exec execpod-affinity45sn9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.23.109 80' +Feb 25 09:51:33.051: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.23.109 80\nConnection to 10.233.23.109 80 port [tcp/http] succeeded!\n" +Feb 25 09:51:33.051: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 09:51:33.051: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6402 exec execpod-affinity45sn9 -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.23.109:80/ ; done' +Feb 25 09:51:33.550: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n" +Feb 25 09:51:33.550: INFO: stdout: "\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8" +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 +Feb 25 09:51:33.551: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6402 exec execpod-affinity45sn9 -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://10.233.23.109:80/' +Feb 25 09:51:33.886: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n" +Feb 25 09:51:33.886: INFO: stdout: "affinity-clusterip-timeout-582x8" +Feb 25 09:51:53.887: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6402 exec execpod-affinity45sn9 -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://10.233.23.109:80/' +Feb 25 09:51:54.176: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n" +Feb 25 09:51:54.177: INFO: stdout: "affinity-clusterip-timeout-t4pcn" +Feb 25 09:51:54.177: INFO: Cleaning up the exec pod +STEP: deleting ReplicationController affinity-clusterip-timeout in namespace services-6402, will wait for the garbage collector to delete the pods 02/25/23 09:51:54.214 +Feb 25 09:51:54.295: INFO: Deleting ReplicationController affinity-clusterip-timeout took: 15.951685ms +Feb 25 09:51:54.396: INFO: Terminating ReplicationController affinity-clusterip-timeout pods took: 101.163702ms [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 15:50:29.088: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-7867" for this suite. 12/14/22 15:50:29.096 +Feb 25 09:51:56.551: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-6402" for this suite. 02/25/23 09:51:56.562 [AfterEach] [sig-network] Services test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance]","completed":24,"skipped":347,"failed":0} +{"msg":"PASSED [sig-network] Services should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance]","completed":23,"skipped":542,"failed":0} ------------------------------ -• [SLOW TEST] [16.142 seconds] +• [SLOW TEST] [44.803 seconds] [sig-network] Services test/e2e/network/common/framework.go:23 - should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance] - test/e2e/network/service.go:2189 + should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance] + test/e2e/network/service.go:2173 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:50:12.965 - Dec 14 15:50:12.966: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 15:50:12.967 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:50:12.993 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:50:13.002 + STEP: Creating a kubernetes client 02/25/23 09:51:11.775 + Feb 25 09:51:11.776: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 09:51:11.787 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:51:11.831 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:51:11.837 [BeforeEach] [sig-network] Services test/e2e/network/service.go:758 - [It] should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance] - test/e2e/network/service.go:2189 - STEP: creating service in namespace services-7867 12/14/22 15:50:13.007 - STEP: creating service affinity-clusterip-transition in namespace services-7867 12/14/22 15:50:13.008 - STEP: creating replication controller affinity-clusterip-transition in namespace services-7867 12/14/22 15:50:13.022 - I1214 15:50:13.033614 14 runners.go:193] Created replication controller with name: affinity-clusterip-transition, namespace: services-7867, replica count: 3 - I1214 15:50:16.085581 14 runners.go:193] affinity-clusterip-transition Pods: 3 out of 3 created, 1 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - I1214 15:50:19.086726 14 runners.go:193] affinity-clusterip-transition Pods: 3 out of 3 created, 1 running, 2 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - I1214 15:50:22.087940 14 runners.go:193] affinity-clusterip-transition Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - Dec 14 15:50:22.103: INFO: Creating new exec pod - Dec 14 15:50:22.119: INFO: Waiting up to 5m0s for pod "execpod-affinityqfh9z" in namespace "services-7867" to be "running" - Dec 14 15:50:22.124: INFO: Pod "execpod-affinityqfh9z": Phase="Pending", Reason="", readiness=false. Elapsed: 5.031003ms - Dec 14 15:50:24.139: INFO: Pod "execpod-affinityqfh9z": Phase="Running", Reason="", readiness=true. Elapsed: 2.020040128s - Dec 14 15:50:24.139: INFO: Pod "execpod-affinityqfh9z" satisfied condition "running" - Dec 14 15:50:25.141: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-7867 exec execpod-affinityqfh9z -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-clusterip-transition 80' - Dec 14 15:50:25.435: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-clusterip-transition 80\nConnection to affinity-clusterip-transition 80 port [tcp/http] succeeded!\n" - Dec 14 15:50:25.435: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 15:50:25.435: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-7867 exec execpod-affinityqfh9z -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.10.125 80' - Dec 14 15:50:25.671: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.10.125 80\nConnection to 10.233.10.125 80 port [tcp/http] succeeded!\n" - Dec 14 15:50:25.671: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 15:50:25.685: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-7867 exec execpod-affinityqfh9z -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.10.125:80/ ; done' - Dec 14 15:50:26.123: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n" - Dec 14 15:50:26.123: INFO: stdout: "\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-zx4lr\naffinity-clusterip-transition-h7xrx\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-zx4lr\naffinity-clusterip-transition-zx4lr\naffinity-clusterip-transition-h7xrx\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-zx4lr\naffinity-clusterip-transition-zx4lr\naffinity-clusterip-transition-h7xrx\naffinity-clusterip-transition-h7xrx\naffinity-clusterip-transition-h7xrx\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl" - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-zx4lr - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-h7xrx - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-zx4lr - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-zx4lr - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-h7xrx - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-zx4lr - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-zx4lr - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-h7xrx - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-h7xrx - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-h7xrx - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.123: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.140: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-7867 exec execpod-affinityqfh9z -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.10.125:80/ ; done' - Dec 14 15:50:26.521: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.10.125:80/\n" - Dec 14 15:50:26.521: INFO: stdout: "\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl\naffinity-clusterip-transition-qjssl" - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Received response from host: affinity-clusterip-transition-qjssl - Dec 14 15:50:26.521: INFO: Cleaning up the exec pod - STEP: deleting ReplicationController affinity-clusterip-transition in namespace services-7867, will wait for the garbage collector to delete the pods 12/14/22 15:50:26.573 - Dec 14 15:50:26.648: INFO: Deleting ReplicationController affinity-clusterip-transition took: 13.726472ms - Dec 14 15:50:26.748: INFO: Terminating ReplicationController affinity-clusterip-transition pods took: 100.590605ms + [It] should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance] + test/e2e/network/service.go:2173 + STEP: creating service in namespace services-6402 02/25/23 09:51:11.849 + Feb 25 09:51:11.868: INFO: Waiting up to 5m0s for pod "kube-proxy-mode-detector" in namespace "services-6402" to be "running and ready" + Feb 25 09:51:11.877: INFO: Pod "kube-proxy-mode-detector": Phase="Pending", Reason="", readiness=false. Elapsed: 8.88539ms + Feb 25 09:51:11.877: INFO: The phase of Pod kube-proxy-mode-detector is Pending, waiting for it to be Running (with Ready = true) + Feb 25 09:51:13.885: INFO: Pod "kube-proxy-mode-detector": Phase="Running", Reason="", readiness=true. Elapsed: 2.016849074s + Feb 25 09:51:13.885: INFO: The phase of Pod kube-proxy-mode-detector is Running (Ready = true) + Feb 25 09:51:13.885: INFO: Pod "kube-proxy-mode-detector" satisfied condition "running and ready" + Feb 25 09:51:13.892: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6402 exec kube-proxy-mode-detector -- /bin/sh -x -c curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode' + Feb 25 09:51:14.215: INFO: stderr: "+ curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode\n" + Feb 25 09:51:14.215: INFO: stdout: "iptables" + Feb 25 09:51:14.215: INFO: proxyMode: iptables + Feb 25 09:51:14.238: INFO: Waiting for pod kube-proxy-mode-detector to disappear + Feb 25 09:51:14.244: INFO: Pod kube-proxy-mode-detector no longer exists + STEP: creating service affinity-clusterip-timeout in namespace services-6402 02/25/23 09:51:14.244 + STEP: creating replication controller affinity-clusterip-timeout in namespace services-6402 02/25/23 09:51:14.263 + I0225 09:51:14.275490 13 runners.go:193] Created replication controller with name: affinity-clusterip-timeout, namespace: services-6402, replica count: 3 + I0225 09:51:17.328761 13 runners.go:193] affinity-clusterip-timeout Pods: 3 out of 3 created, 2 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + I0225 09:51:20.331376 13 runners.go:193] affinity-clusterip-timeout Pods: 3 out of 3 created, 2 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + I0225 09:51:23.333760 13 runners.go:193] affinity-clusterip-timeout Pods: 3 out of 3 created, 2 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + I0225 09:51:26.335445 13 runners.go:193] affinity-clusterip-timeout Pods: 3 out of 3 created, 2 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + I0225 09:51:29.336009 13 runners.go:193] affinity-clusterip-timeout Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + Feb 25 09:51:29.354: INFO: Creating new exec pod + Feb 25 09:51:29.365: INFO: Waiting up to 5m0s for pod "execpod-affinity45sn9" in namespace "services-6402" to be "running" + Feb 25 09:51:29.372: INFO: Pod "execpod-affinity45sn9": Phase="Pending", Reason="", readiness=false. Elapsed: 6.275666ms + Feb 25 09:51:31.382: INFO: Pod "execpod-affinity45sn9": Phase="Running", Reason="", readiness=true. Elapsed: 2.016439057s + Feb 25 09:51:31.382: INFO: Pod "execpod-affinity45sn9" satisfied condition "running" + Feb 25 09:51:32.385: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6402 exec execpod-affinity45sn9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-clusterip-timeout 80' + Feb 25 09:51:32.708: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-clusterip-timeout 80\nConnection to affinity-clusterip-timeout 80 port [tcp/http] succeeded!\n" + Feb 25 09:51:32.708: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 09:51:32.709: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6402 exec execpod-affinity45sn9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.23.109 80' + Feb 25 09:51:33.051: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.23.109 80\nConnection to 10.233.23.109 80 port [tcp/http] succeeded!\n" + Feb 25 09:51:33.051: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 09:51:33.051: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6402 exec execpod-affinity45sn9 -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.23.109:80/ ; done' + Feb 25 09:51:33.550: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n" + Feb 25 09:51:33.550: INFO: stdout: "\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8\naffinity-clusterip-timeout-582x8" + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.550: INFO: Received response from host: affinity-clusterip-timeout-582x8 + Feb 25 09:51:33.551: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6402 exec execpod-affinity45sn9 -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://10.233.23.109:80/' + Feb 25 09:51:33.886: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n" + Feb 25 09:51:33.886: INFO: stdout: "affinity-clusterip-timeout-582x8" + Feb 25 09:51:53.887: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6402 exec execpod-affinity45sn9 -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://10.233.23.109:80/' + Feb 25 09:51:54.176: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://10.233.23.109:80/\n" + Feb 25 09:51:54.177: INFO: stdout: "affinity-clusterip-timeout-t4pcn" + Feb 25 09:51:54.177: INFO: Cleaning up the exec pod + STEP: deleting ReplicationController affinity-clusterip-timeout in namespace services-6402, will wait for the garbage collector to delete the pods 02/25/23 09:51:54.214 + Feb 25 09:51:54.295: INFO: Deleting ReplicationController affinity-clusterip-timeout took: 15.951685ms + Feb 25 09:51:54.396: INFO: Terminating ReplicationController affinity-clusterip-timeout pods took: 101.163702ms [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 15:50:29.088: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-7867" for this suite. 12/14/22 15:50:29.096 + Feb 25 09:51:56.551: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-6402" for this suite. 02/25/23 09:51:56.562 [AfterEach] [sig-network] Services test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] EmptyDir volumes + should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:206 +[BeforeEach] [sig-storage] EmptyDir volumes + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 09:51:56.613 +Feb 25 09:51:56.615: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir 02/25/23 09:51:56.621 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:51:56.661 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:51:56.667 +[It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:206 +STEP: Creating a pod to test emptydir 0666 on node default medium 02/25/23 09:51:56.677 +Feb 25 09:51:56.705: INFO: Waiting up to 5m0s for pod "pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083" in namespace "emptydir-5116" to be "Succeeded or Failed" +Feb 25 09:51:56.711: INFO: Pod "pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083": Phase="Pending", Reason="", readiness=false. Elapsed: 6.442281ms +Feb 25 09:51:58.718: INFO: Pod "pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013451731s +Feb 25 09:52:00.722: INFO: Pod "pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083": Phase="Pending", Reason="", readiness=false. Elapsed: 4.017110556s +Feb 25 09:52:02.719: INFO: Pod "pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.01433957s +STEP: Saw pod success 02/25/23 09:52:02.719 +Feb 25 09:52:02.720: INFO: Pod "pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083" satisfied condition "Succeeded or Failed" +Feb 25 09:52:02.725: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083 container test-container: +STEP: delete the pod 02/25/23 09:52:02.739 +Feb 25 09:52:02.759: INFO: Waiting for pod pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083 to disappear +Feb 25 09:52:02.764: INFO: Pod pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes + test/e2e/framework/framework.go:187 +Feb 25 09:52:02.764: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-5116" for this suite. 02/25/23 09:52:02.77 +{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","completed":24,"skipped":594,"failed":0} +------------------------------ +• [SLOW TEST] [6.170 seconds] +[sig-storage] EmptyDir volumes +test/e2e/common/storage/framework.go:23 + should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:206 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-storage] EmptyDir volumes + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 09:51:56.613 + Feb 25 09:51:56.615: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir 02/25/23 09:51:56.621 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:51:56.661 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:51:56.667 + [It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:206 + STEP: Creating a pod to test emptydir 0666 on node default medium 02/25/23 09:51:56.677 + Feb 25 09:51:56.705: INFO: Waiting up to 5m0s for pod "pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083" in namespace "emptydir-5116" to be "Succeeded or Failed" + Feb 25 09:51:56.711: INFO: Pod "pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083": Phase="Pending", Reason="", readiness=false. Elapsed: 6.442281ms + Feb 25 09:51:58.718: INFO: Pod "pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013451731s + Feb 25 09:52:00.722: INFO: Pod "pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083": Phase="Pending", Reason="", readiness=false. Elapsed: 4.017110556s + Feb 25 09:52:02.719: INFO: Pod "pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.01433957s + STEP: Saw pod success 02/25/23 09:52:02.719 + Feb 25 09:52:02.720: INFO: Pod "pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083" satisfied condition "Succeeded or Failed" + Feb 25 09:52:02.725: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083 container test-container: + STEP: delete the pod 02/25/23 09:52:02.739 + Feb 25 09:52:02.759: INFO: Waiting for pod pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083 to disappear + Feb 25 09:52:02.764: INFO: Pod pod-4a523f9c-9c2d-4a48-8dfd-1155fa526083 no longer exists + [AfterEach] [sig-storage] EmptyDir volumes + test/e2e/framework/framework.go:187 + Feb 25 09:52:02.764: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-5116" for this suite. 02/25/23 09:52:02.77 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-cli] Kubectl client Kubectl run pod + should create a pod from an image when restart is Never [Conformance] + test/e2e/kubectl/kubectl.go:1711 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 09:52:02.808 +Feb 25 09:52:02.808: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 09:52:02.81 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:02.835 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:02.839 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 +[BeforeEach] Kubectl run pod + test/e2e/kubectl/kubectl.go:1698 +[It] should create a pod from an image when restart is Never [Conformance] + test/e2e/kubectl/kubectl.go:1711 +STEP: running the image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 02/25/23 09:52:02.843 +Feb 25 09:52:02.844: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6311 run e2e-test-httpd-pod --restart=Never --pod-running-timeout=2m0s --image=registry.k8s.io/e2e-test-images/httpd:2.4.38-2' +Feb 25 09:52:03.000: INFO: stderr: "" +Feb 25 09:52:03.000: INFO: stdout: "pod/e2e-test-httpd-pod created\n" +STEP: verifying the pod e2e-test-httpd-pod was created 02/25/23 09:52:03 +[AfterEach] Kubectl run pod + test/e2e/kubectl/kubectl.go:1702 +Feb 25 09:52:03.011: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6311 delete pods e2e-test-httpd-pod' +Feb 25 09:52:06.064: INFO: stderr: "" +Feb 25 09:52:06.064: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" +[AfterEach] [sig-cli] Kubectl client + test/e2e/framework/framework.go:187 +Feb 25 09:52:06.064: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-6311" for this suite. 02/25/23 09:52:06.076 +{"msg":"PASSED [sig-cli] Kubectl client Kubectl run pod should create a pod from an image when restart is Never [Conformance]","completed":25,"skipped":732,"failed":0} +------------------------------ +• [3.281 seconds] +[sig-cli] Kubectl client +test/e2e/kubectl/framework.go:23 + Kubectl run pod + test/e2e/kubectl/kubectl.go:1695 + should create a pod from an image when restart is Never [Conformance] + test/e2e/kubectl/kubectl.go:1711 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-cli] Kubectl client + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 09:52:02.808 + Feb 25 09:52:02.808: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 09:52:02.81 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:02.835 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:02.839 + [BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 + [BeforeEach] Kubectl run pod + test/e2e/kubectl/kubectl.go:1698 + [It] should create a pod from an image when restart is Never [Conformance] + test/e2e/kubectl/kubectl.go:1711 + STEP: running the image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 02/25/23 09:52:02.843 + Feb 25 09:52:02.844: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6311 run e2e-test-httpd-pod --restart=Never --pod-running-timeout=2m0s --image=registry.k8s.io/e2e-test-images/httpd:2.4.38-2' + Feb 25 09:52:03.000: INFO: stderr: "" + Feb 25 09:52:03.000: INFO: stdout: "pod/e2e-test-httpd-pod created\n" + STEP: verifying the pod e2e-test-httpd-pod was created 02/25/23 09:52:03 + [AfterEach] Kubectl run pod + test/e2e/kubectl/kubectl.go:1702 + Feb 25 09:52:03.011: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6311 delete pods e2e-test-httpd-pod' + Feb 25 09:52:06.064: INFO: stderr: "" + Feb 25 09:52:06.064: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" + [AfterEach] [sig-cli] Kubectl client + test/e2e/framework/framework.go:187 + Feb 25 09:52:06.064: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-6311" for this suite. 02/25/23 09:52:06.076 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - should mutate custom resource [Conformance] - test/e2e/apimachinery/webhook.go:290 + should honor timeout [Conformance] + test/e2e/apimachinery/webhook.go:380 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:50:29.129 -Dec 14 15:50:29.129: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 15:50:29.132 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:50:29.166 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:50:29.17 +STEP: Creating a kubernetes client 02/25/23 09:52:06.09 +Feb 25 09:52:06.090: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 09:52:06.092 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:06.12 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:06.124 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 15:50:29.197 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 15:50:30.168 -STEP: Deploying the webhook pod 12/14/22 15:50:30.187 -STEP: Wait for the deployment to be ready 12/14/22 15:50:30.213 -Dec 14 15:50:30.240: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set -STEP: Deploying the webhook service 12/14/22 15:50:32.258 -STEP: Verifying the service has paired with the endpoint 12/14/22 15:50:32.274 -Dec 14 15:50:33.275: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] should mutate custom resource [Conformance] - test/e2e/apimachinery/webhook.go:290 -Dec 14 15:50:33.283: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Registering the mutating webhook for custom resource e2e-test-webhook-9980-crds.webhook.example.com via the AdmissionRegistration API 12/14/22 15:50:33.806 -STEP: Creating a custom resource that should be mutated by the webhook 12/14/22 15:50:33.843 +STEP: Setting up server cert 02/25/23 09:52:06.154 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 09:52:07.502 +STEP: Deploying the webhook pod 02/25/23 09:52:07.516 +STEP: Wait for the deployment to be ready 02/25/23 09:52:07.533 +Feb 25 09:52:07.544: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created +STEP: Deploying the webhook service 02/25/23 09:52:09.565 +STEP: Verifying the service has paired with the endpoint 02/25/23 09:52:09.581 +Feb 25 09:52:10.581: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should honor timeout [Conformance] + test/e2e/apimachinery/webhook.go:380 +STEP: Setting timeout (1s) shorter than webhook latency (5s) 02/25/23 09:52:10.588 +STEP: Registering slow webhook via the AdmissionRegistration API 02/25/23 09:52:10.589 +STEP: Request fails when timeout (1s) is shorter than slow webhook latency (5s) 02/25/23 09:52:10.614 +STEP: Having no error when timeout is shorter than webhook latency and failure policy is ignore 02/25/23 09:52:11.645 +STEP: Registering slow webhook via the AdmissionRegistration API 02/25/23 09:52:11.645 +STEP: Having no error when timeout is longer than webhook latency 02/25/23 09:52:12.692 +STEP: Registering slow webhook via the AdmissionRegistration API 02/25/23 09:52:12.692 +STEP: Having no error when timeout is empty (defaulted to 10s in v1) 02/25/23 09:52:17.752 +STEP: Registering slow webhook via the AdmissionRegistration API 02/25/23 09:52:17.752 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 15:50:36.506: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-6686" for this suite. 12/14/22 15:50:36.517 -STEP: Destroying namespace "webhook-6686-markers" for this suite. 12/14/22 15:50:36.526 +Feb 25 09:52:22.809: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-8790" for this suite. 02/25/23 09:52:22.819 +STEP: Destroying namespace "webhook-8790-markers" for this suite. 02/25/23 09:52:22.831 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource [Conformance]","completed":25,"skipped":370,"failed":0} +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should honor timeout [Conformance]","completed":26,"skipped":738,"failed":0} ------------------------------ -• [SLOW TEST] [7.485 seconds] +• [SLOW TEST] [16.851 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/framework.go:23 - should mutate custom resource [Conformance] - test/e2e/apimachinery/webhook.go:290 + should honor timeout [Conformance] + test/e2e/apimachinery/webhook.go:380 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:50:29.129 - Dec 14 15:50:29.129: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 15:50:29.132 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:50:29.166 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:50:29.17 + STEP: Creating a kubernetes client 02/25/23 09:52:06.09 + Feb 25 09:52:06.090: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 09:52:06.092 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:06.12 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:06.124 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 15:50:29.197 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 15:50:30.168 - STEP: Deploying the webhook pod 12/14/22 15:50:30.187 - STEP: Wait for the deployment to be ready 12/14/22 15:50:30.213 - Dec 14 15:50:30.240: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set - STEP: Deploying the webhook service 12/14/22 15:50:32.258 - STEP: Verifying the service has paired with the endpoint 12/14/22 15:50:32.274 - Dec 14 15:50:33.275: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] should mutate custom resource [Conformance] - test/e2e/apimachinery/webhook.go:290 - Dec 14 15:50:33.283: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Registering the mutating webhook for custom resource e2e-test-webhook-9980-crds.webhook.example.com via the AdmissionRegistration API 12/14/22 15:50:33.806 - STEP: Creating a custom resource that should be mutated by the webhook 12/14/22 15:50:33.843 + STEP: Setting up server cert 02/25/23 09:52:06.154 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 09:52:07.502 + STEP: Deploying the webhook pod 02/25/23 09:52:07.516 + STEP: Wait for the deployment to be ready 02/25/23 09:52:07.533 + Feb 25 09:52:07.544: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created + STEP: Deploying the webhook service 02/25/23 09:52:09.565 + STEP: Verifying the service has paired with the endpoint 02/25/23 09:52:09.581 + Feb 25 09:52:10.581: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] should honor timeout [Conformance] + test/e2e/apimachinery/webhook.go:380 + STEP: Setting timeout (1s) shorter than webhook latency (5s) 02/25/23 09:52:10.588 + STEP: Registering slow webhook via the AdmissionRegistration API 02/25/23 09:52:10.589 + STEP: Request fails when timeout (1s) is shorter than slow webhook latency (5s) 02/25/23 09:52:10.614 + STEP: Having no error when timeout is shorter than webhook latency and failure policy is ignore 02/25/23 09:52:11.645 + STEP: Registering slow webhook via the AdmissionRegistration API 02/25/23 09:52:11.645 + STEP: Having no error when timeout is longer than webhook latency 02/25/23 09:52:12.692 + STEP: Registering slow webhook via the AdmissionRegistration API 02/25/23 09:52:12.692 + STEP: Having no error when timeout is empty (defaulted to 10s in v1) 02/25/23 09:52:17.752 + STEP: Registering slow webhook via the AdmissionRegistration API 02/25/23 09:52:17.752 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 15:50:36.506: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-6686" for this suite. 12/14/22 15:50:36.517 - STEP: Destroying namespace "webhook-6686-markers" for this suite. 12/14/22 15:50:36.526 + Feb 25 09:52:22.809: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-8790" for this suite. 02/25/23 09:52:22.819 + STEP: Destroying namespace "webhook-8790-markers" for this suite. 02/25/23 09:52:22.831 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:104 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Projected configMap - should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:46 -[BeforeEach] [sig-storage] Projected configMap +[sig-apps] Daemon set [Serial] + should retry creating failed daemon pods [Conformance] + test/e2e/apps/daemon_set.go:293 +[BeforeEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:50:36.62 -Dec 14 15:50:36.620: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 15:50:36.627 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:50:36.657 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:50:36.662 -[It] should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:46 -STEP: Creating configMap with name projected-configmap-test-volume-7500605c-0e81-4c76-af34-963777f438cb 12/14/22 15:50:36.666 -STEP: Creating a pod to test consume configMaps 12/14/22 15:50:36.677 -Dec 14 15:50:36.694: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1" in namespace "projected-2492" to be "Succeeded or Failed" -Dec 14 15:50:36.703: INFO: Pod "pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1": Phase="Pending", Reason="", readiness=false. Elapsed: 8.266376ms -Dec 14 15:50:38.716: INFO: Pod "pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021231335s -Dec 14 15:50:40.716: INFO: Pod "pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021089334s -STEP: Saw pod success 12/14/22 15:50:40.716 -Dec 14 15:50:40.717: INFO: Pod "pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1" satisfied condition "Succeeded or Failed" -Dec 14 15:50:40.725: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1 container agnhost-container: -STEP: delete the pod 12/14/22 15:50:40.735 -Dec 14 15:50:40.755: INFO: Waiting for pod pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1 to disappear -Dec 14 15:50:40.761: INFO: Pod pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1 no longer exists -[AfterEach] [sig-storage] Projected configMap +STEP: Creating a kubernetes client 02/25/23 09:52:22.986 +Feb 25 09:52:22.987: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename daemonsets 02/25/23 09:52:23.019 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:23.085 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:23.091 +[BeforeEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:145 +[It] should retry creating failed daemon pods [Conformance] + test/e2e/apps/daemon_set.go:293 +STEP: Creating a simple DaemonSet "daemon-set" 02/25/23 09:52:23.138 +STEP: Check that daemon pods launch on every node of the cluster. 02/25/23 09:52:23.151 +Feb 25 09:52:23.164: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 09:52:23.164: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 09:52:24.198: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 09:52:24.198: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 09:52:25.178: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 09:52:25.178: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 09:52:26.181: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 09:52:26.181: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 09:52:27.181: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 09:52:27.181: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 09:52:28.182: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 09:52:28.182: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 09:52:29.193: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 09:52:29.193: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 09:52:30.180: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 09:52:30.180: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 09:52:31.179: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 09:52:31.179: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 09:52:32.179: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 +Feb 25 09:52:32.179: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set +STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived. 02/25/23 09:52:32.185 +Feb 25 09:52:32.231: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 +Feb 25 09:52:32.232: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set +STEP: Wait for the failed daemon pod to be completely deleted. 02/25/23 09:52:32.232 +[AfterEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:110 +STEP: Deleting DaemonSet "daemon-set" 02/25/23 09:52:33.253 +STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-8190, will wait for the garbage collector to delete the pods 02/25/23 09:52:33.253 +Feb 25 09:52:33.329: INFO: Deleting DaemonSet.extensions daemon-set took: 17.71179ms +Feb 25 09:52:33.430: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.983876ms +Feb 25 09:52:34.837: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 09:52:34.837: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set +Feb 25 09:52:34.842: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"10489"},"items":null} + +Feb 25 09:52:34.849: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"10489"},"items":null} + +[AfterEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:187 -Dec 14 15:50:40.761: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-2492" for this suite. 12/14/22 15:50:40.769 -{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance]","completed":26,"skipped":377,"failed":0} +Feb 25 09:52:34.897: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "daemonsets-8190" for this suite. 02/25/23 09:52:34.908 +{"msg":"PASSED [sig-apps] Daemon set [Serial] should retry creating failed daemon pods [Conformance]","completed":27,"skipped":775,"failed":0} ------------------------------ -• [4.164 seconds] -[sig-storage] Projected configMap -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:46 +• [SLOW TEST] [11.937 seconds] +[sig-apps] Daemon set [Serial] +test/e2e/apps/framework.go:23 + should retry creating failed daemon pods [Conformance] + test/e2e/apps/daemon_set.go:293 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected configMap + [BeforeEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:50:36.62 - Dec 14 15:50:36.620: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 15:50:36.627 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:50:36.657 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:50:36.662 - [It] should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:46 - STEP: Creating configMap with name projected-configmap-test-volume-7500605c-0e81-4c76-af34-963777f438cb 12/14/22 15:50:36.666 - STEP: Creating a pod to test consume configMaps 12/14/22 15:50:36.677 - Dec 14 15:50:36.694: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1" in namespace "projected-2492" to be "Succeeded or Failed" - Dec 14 15:50:36.703: INFO: Pod "pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1": Phase="Pending", Reason="", readiness=false. Elapsed: 8.266376ms - Dec 14 15:50:38.716: INFO: Pod "pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021231335s - Dec 14 15:50:40.716: INFO: Pod "pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021089334s - STEP: Saw pod success 12/14/22 15:50:40.716 - Dec 14 15:50:40.717: INFO: Pod "pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1" satisfied condition "Succeeded or Failed" - Dec 14 15:50:40.725: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1 container agnhost-container: - STEP: delete the pod 12/14/22 15:50:40.735 - Dec 14 15:50:40.755: INFO: Waiting for pod pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1 to disappear - Dec 14 15:50:40.761: INFO: Pod pod-projected-configmaps-f0130e52-4931-4549-b186-469e2fce34b1 no longer exists - [AfterEach] [sig-storage] Projected configMap - test/e2e/framework/framework.go:187 - Dec 14 15:50:40.761: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-2492" for this suite. 12/14/22 15:50:40.769 - << End Captured GinkgoWriter Output ------------------------------- -SSSSS + STEP: Creating a kubernetes client 02/25/23 09:52:22.986 + Feb 25 09:52:22.987: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename daemonsets 02/25/23 09:52:23.019 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:23.085 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:23.091 + [BeforeEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:145 + [It] should retry creating failed daemon pods [Conformance] + test/e2e/apps/daemon_set.go:293 + STEP: Creating a simple DaemonSet "daemon-set" 02/25/23 09:52:23.138 + STEP: Check that daemon pods launch on every node of the cluster. 02/25/23 09:52:23.151 + Feb 25 09:52:23.164: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 09:52:23.164: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 09:52:24.198: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 09:52:24.198: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 09:52:25.178: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 09:52:25.178: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 09:52:26.181: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 09:52:26.181: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 09:52:27.181: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 09:52:27.181: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 09:52:28.182: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 09:52:28.182: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 09:52:29.193: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 09:52:29.193: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 09:52:30.180: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 09:52:30.180: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 09:52:31.179: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 09:52:31.179: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 09:52:32.179: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 + Feb 25 09:52:32.179: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set + STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived. 02/25/23 09:52:32.185 + Feb 25 09:52:32.231: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 + Feb 25 09:52:32.232: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set + STEP: Wait for the failed daemon pod to be completely deleted. 02/25/23 09:52:32.232 + [AfterEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:110 + STEP: Deleting DaemonSet "daemon-set" 02/25/23 09:52:33.253 + STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-8190, will wait for the garbage collector to delete the pods 02/25/23 09:52:33.253 + Feb 25 09:52:33.329: INFO: Deleting DaemonSet.extensions daemon-set took: 17.71179ms + Feb 25 09:52:33.430: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.983876ms + Feb 25 09:52:34.837: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 09:52:34.837: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set + Feb 25 09:52:34.842: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"10489"},"items":null} + + Feb 25 09:52:34.849: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"10489"},"items":null} + + [AfterEach] [sig-apps] Daemon set [Serial] + test/e2e/framework/framework.go:187 + Feb 25 09:52:34.897: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "daemonsets-8190" for this suite. 02/25/23 09:52:34.908 + << End Captured GinkgoWriter Output ------------------------------ -[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] - should perform canary updates and phased rolling updates of template modifications [Conformance] - test/e2e/apps/statefulset.go:315 -[BeforeEach] [sig-apps] StatefulSet +SSSSS +------------------------------ +[sig-api-machinery] ResourceQuota + should manage the lifecycle of a ResourceQuota [Conformance] + test/e2e/apimachinery/resource_quota.go:933 +[BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:50:40.791 -Dec 14 15:50:40.792: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename statefulset 12/14/22 15:50:40.795 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:50:40.826 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:50:40.831 -[BeforeEach] [sig-apps] StatefulSet - test/e2e/apps/statefulset.go:96 -[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:111 -STEP: Creating service test in namespace statefulset-8032 12/14/22 15:50:40.837 -[It] should perform canary updates and phased rolling updates of template modifications [Conformance] - test/e2e/apps/statefulset.go:315 -STEP: Creating a new StatefulSet 12/14/22 15:50:40.848 -Dec 14 15:50:40.871: INFO: Found 0 stateful pods, waiting for 3 -Dec 14 15:50:50.882: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true -Dec 14 15:50:50.882: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true -Dec 14 15:50:50.882: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true -STEP: Updating stateful set template: update image from registry.k8s.io/e2e-test-images/httpd:2.4.38-2 to registry.k8s.io/e2e-test-images/httpd:2.4.39-2 12/14/22 15:50:50.907 -Dec 14 15:50:50.936: INFO: Updating stateful set ss2 -STEP: Creating a new revision 12/14/22 15:50:50.936 -STEP: Not applying an update when the partition is greater than the number of replicas 12/14/22 15:51:00.975 -STEP: Performing a canary update 12/14/22 15:51:00.975 -Dec 14 15:51:01.009: INFO: Updating stateful set ss2 -Dec 14 15:51:01.020: INFO: Waiting for Pod statefulset-8032/ss2-2 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 -STEP: Restoring Pods to the correct revision when they are deleted 12/14/22 15:51:11.035 -Dec 14 15:51:11.125: INFO: Found 1 stateful pods, waiting for 3 -Dec 14 15:51:21.142: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true -Dec 14 15:51:21.142: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true -Dec 14 15:51:21.142: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true -STEP: Performing a phased rolling update 12/14/22 15:51:21.16 -Dec 14 15:51:21.197: INFO: Updating stateful set ss2 -Dec 14 15:51:21.215: INFO: Waiting for Pod statefulset-8032/ss2-1 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 -Dec 14 15:51:31.270: INFO: Updating stateful set ss2 -Dec 14 15:51:31.285: INFO: Waiting for StatefulSet statefulset-8032/ss2 to complete update -Dec 14 15:51:31.285: INFO: Waiting for Pod statefulset-8032/ss2-0 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 -[AfterEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:122 -Dec 14 15:51:41.312: INFO: Deleting all statefulset in ns statefulset-8032 -Dec 14 15:51:41.317: INFO: Scaling statefulset ss2 to 0 -Dec 14 15:51:51.351: INFO: Waiting for statefulset status.replicas updated to 0 -Dec 14 15:51:51.359: INFO: Deleting statefulset ss2 -[AfterEach] [sig-apps] StatefulSet +STEP: Creating a kubernetes client 02/25/23 09:52:34.928 +Feb 25 09:52:34.929: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename resourcequota 02/25/23 09:52:34.932 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:34.962 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:34.968 +[It] should manage the lifecycle of a ResourceQuota [Conformance] + test/e2e/apimachinery/resource_quota.go:933 +STEP: Creating a ResourceQuota 02/25/23 09:52:34.971 +STEP: Getting a ResourceQuota 02/25/23 09:52:34.978 +STEP: Listing all ResourceQuotas with LabelSelector 02/25/23 09:52:34.984 +STEP: Patching the ResourceQuota 02/25/23 09:52:34.993 +STEP: Deleting a Collection of ResourceQuotas 02/25/23 09:52:35.006 +STEP: Verifying the deleted ResourceQuota 02/25/23 09:52:35.029 +[AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 -Dec 14 15:51:51.377: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "statefulset-8032" for this suite. 12/14/22 15:51:51.383 -{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance]","completed":27,"skipped":382,"failed":0} +Feb 25 09:52:35.036: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-3008" for this suite. 02/25/23 09:52:35.054 +{"msg":"PASSED [sig-api-machinery] ResourceQuota should manage the lifecycle of a ResourceQuota [Conformance]","completed":28,"skipped":780,"failed":0} ------------------------------ -• [SLOW TEST] [70.600 seconds] -[sig-apps] StatefulSet -test/e2e/apps/framework.go:23 - Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:101 - should perform canary updates and phased rolling updates of template modifications [Conformance] - test/e2e/apps/statefulset.go:315 +• [0.138 seconds] +[sig-api-machinery] ResourceQuota +test/e2e/apimachinery/framework.go:23 + should manage the lifecycle of a ResourceQuota [Conformance] + test/e2e/apimachinery/resource_quota.go:933 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] StatefulSet + [BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:50:40.791 - Dec 14 15:50:40.792: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename statefulset 12/14/22 15:50:40.795 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:50:40.826 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:50:40.831 - [BeforeEach] [sig-apps] StatefulSet - test/e2e/apps/statefulset.go:96 - [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:111 - STEP: Creating service test in namespace statefulset-8032 12/14/22 15:50:40.837 - [It] should perform canary updates and phased rolling updates of template modifications [Conformance] - test/e2e/apps/statefulset.go:315 - STEP: Creating a new StatefulSet 12/14/22 15:50:40.848 - Dec 14 15:50:40.871: INFO: Found 0 stateful pods, waiting for 3 - Dec 14 15:50:50.882: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true - Dec 14 15:50:50.882: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true - Dec 14 15:50:50.882: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true - STEP: Updating stateful set template: update image from registry.k8s.io/e2e-test-images/httpd:2.4.38-2 to registry.k8s.io/e2e-test-images/httpd:2.4.39-2 12/14/22 15:50:50.907 - Dec 14 15:50:50.936: INFO: Updating stateful set ss2 - STEP: Creating a new revision 12/14/22 15:50:50.936 - STEP: Not applying an update when the partition is greater than the number of replicas 12/14/22 15:51:00.975 - STEP: Performing a canary update 12/14/22 15:51:00.975 - Dec 14 15:51:01.009: INFO: Updating stateful set ss2 - Dec 14 15:51:01.020: INFO: Waiting for Pod statefulset-8032/ss2-2 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 - STEP: Restoring Pods to the correct revision when they are deleted 12/14/22 15:51:11.035 - Dec 14 15:51:11.125: INFO: Found 1 stateful pods, waiting for 3 - Dec 14 15:51:21.142: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true - Dec 14 15:51:21.142: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true - Dec 14 15:51:21.142: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true - STEP: Performing a phased rolling update 12/14/22 15:51:21.16 - Dec 14 15:51:21.197: INFO: Updating stateful set ss2 - Dec 14 15:51:21.215: INFO: Waiting for Pod statefulset-8032/ss2-1 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 - Dec 14 15:51:31.270: INFO: Updating stateful set ss2 - Dec 14 15:51:31.285: INFO: Waiting for StatefulSet statefulset-8032/ss2 to complete update - Dec 14 15:51:31.285: INFO: Waiting for Pod statefulset-8032/ss2-0 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 - [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:122 - Dec 14 15:51:41.312: INFO: Deleting all statefulset in ns statefulset-8032 - Dec 14 15:51:41.317: INFO: Scaling statefulset ss2 to 0 - Dec 14 15:51:51.351: INFO: Waiting for statefulset status.replicas updated to 0 - Dec 14 15:51:51.359: INFO: Deleting statefulset ss2 - [AfterEach] [sig-apps] StatefulSet + STEP: Creating a kubernetes client 02/25/23 09:52:34.928 + Feb 25 09:52:34.929: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename resourcequota 02/25/23 09:52:34.932 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:34.962 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:34.968 + [It] should manage the lifecycle of a ResourceQuota [Conformance] + test/e2e/apimachinery/resource_quota.go:933 + STEP: Creating a ResourceQuota 02/25/23 09:52:34.971 + STEP: Getting a ResourceQuota 02/25/23 09:52:34.978 + STEP: Listing all ResourceQuotas with LabelSelector 02/25/23 09:52:34.984 + STEP: Patching the ResourceQuota 02/25/23 09:52:34.993 + STEP: Deleting a Collection of ResourceQuotas 02/25/23 09:52:35.006 + STEP: Verifying the deleted ResourceQuota 02/25/23 09:52:35.029 + [AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 - Dec 14 15:51:51.377: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "statefulset-8032" for this suite. 12/14/22 15:51:51.383 + Feb 25 09:52:35.036: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "resourcequota-3008" for this suite. 02/25/23 09:52:35.054 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Variable Expansion - should succeed in writing subpaths in container [Slow] [Conformance] - test/e2e/common/node/expansion.go:296 -[BeforeEach] [sig-node] Variable Expansion +[sig-network] EndpointSlice + should create and delete Endpoints and EndpointSlices for a Service with a selector specified [Conformance] + test/e2e/network/endpointslice.go:101 +[BeforeEach] [sig-network] EndpointSlice test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:51:51.396 -Dec 14 15:51:51.396: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename var-expansion 12/14/22 15:51:51.399 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:51:51.426 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:51:51.429 -[It] should succeed in writing subpaths in container [Slow] [Conformance] - test/e2e/common/node/expansion.go:296 -STEP: creating the pod 12/14/22 15:51:51.433 -STEP: waiting for pod running 12/14/22 15:51:51.444 -Dec 14 15:51:51.444: INFO: Waiting up to 2m0s for pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2" in namespace "var-expansion-3060" to be "running" -Dec 14 15:51:51.448: INFO: Pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2": Phase="Pending", Reason="", readiness=false. Elapsed: 3.899129ms -Dec 14 15:51:53.458: INFO: Pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2": Phase="Running", Reason="", readiness=true. Elapsed: 2.013920656s -Dec 14 15:51:53.459: INFO: Pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2" satisfied condition "running" -STEP: creating a file in subpath 12/14/22 15:51:53.459 -Dec 14 15:51:53.465: INFO: ExecWithOptions {Command:[/bin/sh -c touch /volume_mount/mypath/foo/test.log] Namespace:var-expansion-3060 PodName:var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2 ContainerName:dapi-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 15:51:53.465: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 15:51:53.468: INFO: ExecWithOptions: Clientset creation -Dec 14 15:51:53.468: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/var-expansion-3060/pods/var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2/exec?command=%2Fbin%2Fsh&command=-c&command=touch+%2Fvolume_mount%2Fmypath%2Ffoo%2Ftest.log&container=dapi-container&container=dapi-container&stderr=true&stdout=true) -STEP: test for file in mounted path 12/14/22 15:51:53.551 -Dec 14 15:51:53.558: INFO: ExecWithOptions {Command:[/bin/sh -c test -f /subpath_mount/test.log] Namespace:var-expansion-3060 PodName:var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2 ContainerName:dapi-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 15:51:53.558: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 15:51:53.559: INFO: ExecWithOptions: Clientset creation -Dec 14 15:51:53.559: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/var-expansion-3060/pods/var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2/exec?command=%2Fbin%2Fsh&command=-c&command=test+-f+%2Fsubpath_mount%2Ftest.log&container=dapi-container&container=dapi-container&stderr=true&stdout=true) -STEP: updating the annotation value 12/14/22 15:51:53.632 -Dec 14 15:51:54.153: INFO: Successfully updated pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2" -STEP: waiting for annotated pod running 12/14/22 15:51:54.153 -Dec 14 15:51:54.154: INFO: Waiting up to 2m0s for pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2" in namespace "var-expansion-3060" to be "running" -Dec 14 15:51:54.158: INFO: Pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2": Phase="Running", Reason="", readiness=true. Elapsed: 4.449313ms -Dec 14 15:51:54.158: INFO: Pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2" satisfied condition "running" -STEP: deleting the pod gracefully 12/14/22 15:51:54.159 -Dec 14 15:51:54.159: INFO: Deleting pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2" in namespace "var-expansion-3060" -Dec 14 15:51:54.169: INFO: Wait up to 5m0s for pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2" to be fully deleted -[AfterEach] [sig-node] Variable Expansion +STEP: Creating a kubernetes client 02/25/23 09:52:35.08 +Feb 25 09:52:35.081: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename endpointslice 02/25/23 09:52:35.083 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:35.118 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:35.123 +[BeforeEach] [sig-network] EndpointSlice + test/e2e/network/endpointslice.go:51 +[It] should create and delete Endpoints and EndpointSlices for a Service with a selector specified [Conformance] + test/e2e/network/endpointslice.go:101 +[AfterEach] [sig-network] EndpointSlice test/e2e/framework/framework.go:187 -Dec 14 15:52:28.186: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "var-expansion-3060" for this suite. 12/14/22 15:52:28.195 -{"msg":"PASSED [sig-node] Variable Expansion should succeed in writing subpaths in container [Slow] [Conformance]","completed":28,"skipped":393,"failed":0} +Feb 25 09:52:35.215: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "endpointslice-5844" for this suite. 02/25/23 09:52:35.225 +{"msg":"PASSED [sig-network] EndpointSlice should create and delete Endpoints and EndpointSlices for a Service with a selector specified [Conformance]","completed":29,"skipped":810,"failed":0} ------------------------------ -• [SLOW TEST] [36.814 seconds] -[sig-node] Variable Expansion -test/e2e/common/node/framework.go:23 - should succeed in writing subpaths in container [Slow] [Conformance] - test/e2e/common/node/expansion.go:296 +• [0.160 seconds] +[sig-network] EndpointSlice +test/e2e/network/common/framework.go:23 + should create and delete Endpoints and EndpointSlices for a Service with a selector specified [Conformance] + test/e2e/network/endpointslice.go:101 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Variable Expansion + [BeforeEach] [sig-network] EndpointSlice test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:51:51.396 - Dec 14 15:51:51.396: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename var-expansion 12/14/22 15:51:51.399 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:51:51.426 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:51:51.429 - [It] should succeed in writing subpaths in container [Slow] [Conformance] - test/e2e/common/node/expansion.go:296 - STEP: creating the pod 12/14/22 15:51:51.433 - STEP: waiting for pod running 12/14/22 15:51:51.444 - Dec 14 15:51:51.444: INFO: Waiting up to 2m0s for pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2" in namespace "var-expansion-3060" to be "running" - Dec 14 15:51:51.448: INFO: Pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2": Phase="Pending", Reason="", readiness=false. Elapsed: 3.899129ms - Dec 14 15:51:53.458: INFO: Pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2": Phase="Running", Reason="", readiness=true. Elapsed: 2.013920656s - Dec 14 15:51:53.459: INFO: Pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2" satisfied condition "running" - STEP: creating a file in subpath 12/14/22 15:51:53.459 - Dec 14 15:51:53.465: INFO: ExecWithOptions {Command:[/bin/sh -c touch /volume_mount/mypath/foo/test.log] Namespace:var-expansion-3060 PodName:var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2 ContainerName:dapi-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 15:51:53.465: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 15:51:53.468: INFO: ExecWithOptions: Clientset creation - Dec 14 15:51:53.468: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/var-expansion-3060/pods/var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2/exec?command=%2Fbin%2Fsh&command=-c&command=touch+%2Fvolume_mount%2Fmypath%2Ffoo%2Ftest.log&container=dapi-container&container=dapi-container&stderr=true&stdout=true) - STEP: test for file in mounted path 12/14/22 15:51:53.551 - Dec 14 15:51:53.558: INFO: ExecWithOptions {Command:[/bin/sh -c test -f /subpath_mount/test.log] Namespace:var-expansion-3060 PodName:var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2 ContainerName:dapi-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 15:51:53.558: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 15:51:53.559: INFO: ExecWithOptions: Clientset creation - Dec 14 15:51:53.559: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/var-expansion-3060/pods/var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2/exec?command=%2Fbin%2Fsh&command=-c&command=test+-f+%2Fsubpath_mount%2Ftest.log&container=dapi-container&container=dapi-container&stderr=true&stdout=true) - STEP: updating the annotation value 12/14/22 15:51:53.632 - Dec 14 15:51:54.153: INFO: Successfully updated pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2" - STEP: waiting for annotated pod running 12/14/22 15:51:54.153 - Dec 14 15:51:54.154: INFO: Waiting up to 2m0s for pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2" in namespace "var-expansion-3060" to be "running" - Dec 14 15:51:54.158: INFO: Pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2": Phase="Running", Reason="", readiness=true. Elapsed: 4.449313ms - Dec 14 15:51:54.158: INFO: Pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2" satisfied condition "running" - STEP: deleting the pod gracefully 12/14/22 15:51:54.159 - Dec 14 15:51:54.159: INFO: Deleting pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2" in namespace "var-expansion-3060" - Dec 14 15:51:54.169: INFO: Wait up to 5m0s for pod "var-expansion-da0ccdb5-289d-47b9-bd69-4b963e1485a2" to be fully deleted - [AfterEach] [sig-node] Variable Expansion + STEP: Creating a kubernetes client 02/25/23 09:52:35.08 + Feb 25 09:52:35.081: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename endpointslice 02/25/23 09:52:35.083 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:35.118 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:35.123 + [BeforeEach] [sig-network] EndpointSlice + test/e2e/network/endpointslice.go:51 + [It] should create and delete Endpoints and EndpointSlices for a Service with a selector specified [Conformance] + test/e2e/network/endpointslice.go:101 + [AfterEach] [sig-network] EndpointSlice test/e2e/framework/framework.go:187 - Dec 14 15:52:28.186: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "var-expansion-3060" for this suite. 12/14/22 15:52:28.195 + Feb 25 09:52:35.215: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "endpointslice-5844" for this suite. 02/25/23 09:52:35.225 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSS ------------------------------ -[sig-apps] Deployment - RollingUpdateDeployment should delete old pods and create new ones [Conformance] - test/e2e/apps/deployment.go:105 -[BeforeEach] [sig-apps] Deployment +[sig-storage] Secrets + should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:67 +[BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:52:28.227 -Dec 14 15:52:28.227: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename deployment 12/14/22 15:52:28.23 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:52:28.259 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:52:28.264 -[BeforeEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:91 -[It] RollingUpdateDeployment should delete old pods and create new ones [Conformance] - test/e2e/apps/deployment.go:105 -Dec 14 15:52:28.267: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted) -Dec 14 15:52:28.281: INFO: Pod name sample-pod: Found 0 pods out of 1 -Dec 14 15:52:33.290: INFO: Pod name sample-pod: Found 1 pods out of 1 -STEP: ensuring each pod is running 12/14/22 15:52:33.29 -Dec 14 15:52:33.291: INFO: Creating deployment "test-rolling-update-deployment" -Dec 14 15:52:33.301: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has -Dec 14 15:52:33.310: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created -Dec 14 15:52:35.333: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected -Dec 14 15:52:35.341: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted) -[AfterEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:84 -Dec 14 15:52:35.358: INFO: Deployment "test-rolling-update-deployment": -&Deployment{ObjectMeta:{test-rolling-update-deployment deployment-9163 5af465e7-4c7f-4825-8c3a-4ca18a3eacca 5564 1 2022-12-14 15:52:33 +0000 UTC map[name:sample-pod] map[deployment.kubernetes.io/revision:3546343826724305833] [] [] [{e2e.test Update apps/v1 2022-12-14 15:52:33 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 15:52:34 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0025bafe8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2022-12-14 15:52:33 +0000 UTC,LastTransitionTime:2022-12-14 15:52:33 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rolling-update-deployment-78f575d8ff" has successfully progressed.,LastUpdateTime:2022-12-14 15:52:34 +0000 UTC,LastTransitionTime:2022-12-14 15:52:33 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} - -Dec 14 15:52:35.362: INFO: New ReplicaSet "test-rolling-update-deployment-78f575d8ff" of Deployment "test-rolling-update-deployment": -&ReplicaSet{ObjectMeta:{test-rolling-update-deployment-78f575d8ff deployment-9163 df2ed085-cd09-4b11-8e11-f5591317649c 5554 1 2022-12-14 15:52:33 +0000 UTC map[name:sample-pod pod-template-hash:78f575d8ff] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305833] [{apps/v1 Deployment test-rolling-update-deployment 5af465e7-4c7f-4825-8c3a-4ca18a3eacca 0xc0025bb4c7 0xc0025bb4c8}] [] [{kube-controller-manager Update apps/v1 2022-12-14 15:52:33 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"5af465e7-4c7f-4825-8c3a-4ca18a3eacca\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 15:52:34 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 78f575d8ff,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod-template-hash:78f575d8ff] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0025bb578 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} -Dec 14 15:52:35.363: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment": -Dec 14 15:52:35.363: INFO: &ReplicaSet{ObjectMeta:{test-rolling-update-controller deployment-9163 57125f12-0e5a-4837-8ef0-e848e750a8b3 5563 2 2022-12-14 15:52:28 +0000 UTC map[name:sample-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305832] [{apps/v1 Deployment test-rolling-update-deployment 5af465e7-4c7f-4825-8c3a-4ca18a3eacca 0xc0025bb39f 0xc0025bb3b0}] [] [{e2e.test Update apps/v1 2022-12-14 15:52:28 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 15:52:34 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"5af465e7-4c7f-4825-8c3a-4ca18a3eacca\"}":{}}},"f:spec":{"f:replicas":{}}} } {kube-controller-manager Update apps/v1 2022-12-14 15:52:34 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc0025bb468 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} -Dec 14 15:52:35.371: INFO: Pod "test-rolling-update-deployment-78f575d8ff-h98sr" is available: -&Pod{ObjectMeta:{test-rolling-update-deployment-78f575d8ff-h98sr test-rolling-update-deployment-78f575d8ff- deployment-9163 698a4db4-ae5a-4ebb-9430-e886424f02f9 5553 0 2022-12-14 15:52:33 +0000 UTC map[name:sample-pod pod-template-hash:78f575d8ff] map[] [{apps/v1 ReplicaSet test-rolling-update-deployment-78f575d8ff df2ed085-cd09-4b11-8e11-f5591317649c 0xc003681567 0xc003681568}] [] [{kube-controller-manager Update v1 2022-12-14 15:52:33 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"df2ed085-cd09-4b11-8e11-f5591317649c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 15:52:34 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.39\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-wxgl2,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wxgl2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:52:33 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:52:34 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:52:34 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:52:33 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.39,StartTime:2022-12-14 15:52:33 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 15:52:34 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,ImageID:registry.k8s.io/e2e-test-images/agnhost@sha256:37dd9ad84b34913b4bc57e00b0a28e8b00f18ca6a5ecec69461aa57402e5c787,ContainerID:cri-o://26fc5aff74ebcd80c8171b68fc37749bd384ff0dcff61fff0888e5b2cc4131c8,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.39,},},EphemeralContainerStatuses:[]ContainerStatus{},},} -[AfterEach] [sig-apps] Deployment +STEP: Creating a kubernetes client 02/25/23 09:52:35.25 +Feb 25 09:52:35.250: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename secrets 02/25/23 09:52:35.258 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:35.291 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:35.295 +[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:67 +STEP: Creating secret with name secret-test-33464230-7a24-4a3f-9f10-45395492586c 02/25/23 09:52:35.3 +STEP: Creating a pod to test consume secrets 02/25/23 09:52:35.309 +Feb 25 09:52:35.324: INFO: Waiting up to 5m0s for pod "pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6" in namespace "secrets-2125" to be "Succeeded or Failed" +Feb 25 09:52:35.330: INFO: Pod "pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6": Phase="Pending", Reason="", readiness=false. Elapsed: 6.05957ms +Feb 25 09:52:37.338: INFO: Pod "pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013947888s +Feb 25 09:52:39.337: INFO: Pod "pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013012517s +STEP: Saw pod success 02/25/23 09:52:39.338 +Feb 25 09:52:39.338: INFO: Pod "pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6" satisfied condition "Succeeded or Failed" +Feb 25 09:52:39.343: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6 container secret-volume-test: +STEP: delete the pod 02/25/23 09:52:39.357 +Feb 25 09:52:39.376: INFO: Waiting for pod pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6 to disappear +Feb 25 09:52:39.381: INFO: Pod pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6 no longer exists +[AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 -Dec 14 15:52:35.371: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "deployment-9163" for this suite. 12/14/22 15:52:35.379 -{"msg":"PASSED [sig-apps] Deployment RollingUpdateDeployment should delete old pods and create new ones [Conformance]","completed":29,"skipped":445,"failed":0} +Feb 25 09:52:39.381: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-2125" for this suite. 02/25/23 09:52:39.387 +{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","completed":30,"skipped":824,"failed":0} ------------------------------ -• [SLOW TEST] [7.164 seconds] -[sig-apps] Deployment -test/e2e/apps/framework.go:23 - RollingUpdateDeployment should delete old pods and create new ones [Conformance] - test/e2e/apps/deployment.go:105 +• [4.147 seconds] +[sig-storage] Secrets +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:67 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Deployment + [BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:52:28.227 - Dec 14 15:52:28.227: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename deployment 12/14/22 15:52:28.23 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:52:28.259 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:52:28.264 - [BeforeEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:91 - [It] RollingUpdateDeployment should delete old pods and create new ones [Conformance] - test/e2e/apps/deployment.go:105 - Dec 14 15:52:28.267: INFO: Creating replica set "test-rolling-update-controller" (going to be adopted) - Dec 14 15:52:28.281: INFO: Pod name sample-pod: Found 0 pods out of 1 - Dec 14 15:52:33.290: INFO: Pod name sample-pod: Found 1 pods out of 1 - STEP: ensuring each pod is running 12/14/22 15:52:33.29 - Dec 14 15:52:33.291: INFO: Creating deployment "test-rolling-update-deployment" - Dec 14 15:52:33.301: INFO: Ensuring deployment "test-rolling-update-deployment" gets the next revision from the one the adopted replica set "test-rolling-update-controller" has - Dec 14 15:52:33.310: INFO: new replicaset for deployment "test-rolling-update-deployment" is yet to be created - Dec 14 15:52:35.333: INFO: Ensuring status for deployment "test-rolling-update-deployment" is the expected - Dec 14 15:52:35.341: INFO: Ensuring deployment "test-rolling-update-deployment" has one old replica set (the one it adopted) - [AfterEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:84 - Dec 14 15:52:35.358: INFO: Deployment "test-rolling-update-deployment": - &Deployment{ObjectMeta:{test-rolling-update-deployment deployment-9163 5af465e7-4c7f-4825-8c3a-4ca18a3eacca 5564 1 2022-12-14 15:52:33 +0000 UTC map[name:sample-pod] map[deployment.kubernetes.io/revision:3546343826724305833] [] [] [{e2e.test Update apps/v1 2022-12-14 15:52:33 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 15:52:34 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0025bafe8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2022-12-14 15:52:33 +0000 UTC,LastTransitionTime:2022-12-14 15:52:33 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rolling-update-deployment-78f575d8ff" has successfully progressed.,LastUpdateTime:2022-12-14 15:52:34 +0000 UTC,LastTransitionTime:2022-12-14 15:52:33 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} - - Dec 14 15:52:35.362: INFO: New ReplicaSet "test-rolling-update-deployment-78f575d8ff" of Deployment "test-rolling-update-deployment": - &ReplicaSet{ObjectMeta:{test-rolling-update-deployment-78f575d8ff deployment-9163 df2ed085-cd09-4b11-8e11-f5591317649c 5554 1 2022-12-14 15:52:33 +0000 UTC map[name:sample-pod pod-template-hash:78f575d8ff] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305833] [{apps/v1 Deployment test-rolling-update-deployment 5af465e7-4c7f-4825-8c3a-4ca18a3eacca 0xc0025bb4c7 0xc0025bb4c8}] [] [{kube-controller-manager Update apps/v1 2022-12-14 15:52:33 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"5af465e7-4c7f-4825-8c3a-4ca18a3eacca\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 15:52:34 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod-template-hash: 78f575d8ff,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod-template-hash:78f575d8ff] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0025bb578 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} - Dec 14 15:52:35.363: INFO: All old ReplicaSets of Deployment "test-rolling-update-deployment": - Dec 14 15:52:35.363: INFO: &ReplicaSet{ObjectMeta:{test-rolling-update-controller deployment-9163 57125f12-0e5a-4837-8ef0-e848e750a8b3 5563 2 2022-12-14 15:52:28 +0000 UTC map[name:sample-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:3546343826724305832] [{apps/v1 Deployment test-rolling-update-deployment 5af465e7-4c7f-4825-8c3a-4ca18a3eacca 0xc0025bb39f 0xc0025bb3b0}] [] [{e2e.test Update apps/v1 2022-12-14 15:52:28 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 15:52:34 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"5af465e7-4c7f-4825-8c3a-4ca18a3eacca\"}":{}}},"f:spec":{"f:replicas":{}}} } {kube-controller-manager Update apps/v1 2022-12-14 15:52:34 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod pod:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc0025bb468 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} - Dec 14 15:52:35.371: INFO: Pod "test-rolling-update-deployment-78f575d8ff-h98sr" is available: - &Pod{ObjectMeta:{test-rolling-update-deployment-78f575d8ff-h98sr test-rolling-update-deployment-78f575d8ff- deployment-9163 698a4db4-ae5a-4ebb-9430-e886424f02f9 5553 0 2022-12-14 15:52:33 +0000 UTC map[name:sample-pod pod-template-hash:78f575d8ff] map[] [{apps/v1 ReplicaSet test-rolling-update-deployment-78f575d8ff df2ed085-cd09-4b11-8e11-f5591317649c 0xc003681567 0xc003681568}] [] [{kube-controller-manager Update v1 2022-12-14 15:52:33 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"df2ed085-cd09-4b11-8e11-f5591317649c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 15:52:34 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.39\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-wxgl2,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wxgl2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:52:33 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:52:34 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:52:34 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:52:33 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.39,StartTime:2022-12-14 15:52:33 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 15:52:34 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,ImageID:registry.k8s.io/e2e-test-images/agnhost@sha256:37dd9ad84b34913b4bc57e00b0a28e8b00f18ca6a5ecec69461aa57402e5c787,ContainerID:cri-o://26fc5aff74ebcd80c8171b68fc37749bd384ff0dcff61fff0888e5b2cc4131c8,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.39,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - [AfterEach] [sig-apps] Deployment + STEP: Creating a kubernetes client 02/25/23 09:52:35.25 + Feb 25 09:52:35.250: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename secrets 02/25/23 09:52:35.258 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:35.291 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:35.295 + [It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:67 + STEP: Creating secret with name secret-test-33464230-7a24-4a3f-9f10-45395492586c 02/25/23 09:52:35.3 + STEP: Creating a pod to test consume secrets 02/25/23 09:52:35.309 + Feb 25 09:52:35.324: INFO: Waiting up to 5m0s for pod "pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6" in namespace "secrets-2125" to be "Succeeded or Failed" + Feb 25 09:52:35.330: INFO: Pod "pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6": Phase="Pending", Reason="", readiness=false. Elapsed: 6.05957ms + Feb 25 09:52:37.338: INFO: Pod "pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013947888s + Feb 25 09:52:39.337: INFO: Pod "pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013012517s + STEP: Saw pod success 02/25/23 09:52:39.338 + Feb 25 09:52:39.338: INFO: Pod "pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6" satisfied condition "Succeeded or Failed" + Feb 25 09:52:39.343: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6 container secret-volume-test: + STEP: delete the pod 02/25/23 09:52:39.357 + Feb 25 09:52:39.376: INFO: Waiting for pod pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6 to disappear + Feb 25 09:52:39.381: INFO: Pod pod-secrets-fcf2f84c-7889-4c21-9415-7b98656480a6 no longer exists + [AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 - Dec 14 15:52:35.371: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "deployment-9163" for this suite. 12/14/22 15:52:35.379 + Feb 25 09:52:39.381: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "secrets-2125" for this suite. 02/25/23 09:52:39.387 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Projected downwardAPI - should provide container's memory request [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:234 -[BeforeEach] [sig-storage] Projected downwardAPI +[sig-storage] Projected secret + should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:66 +[BeforeEach] [sig-storage] Projected secret test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:52:35.399 -Dec 14 15:52:35.399: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 15:52:35.403 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:52:35.437 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:52:35.441 -[BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 -[It] should provide container's memory request [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:234 -STEP: Creating a pod to test downward API volume plugin 12/14/22 15:52:35.444 -Dec 14 15:52:35.460: INFO: Waiting up to 5m0s for pod "downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5" in namespace "projected-2948" to be "Succeeded or Failed" -Dec 14 15:52:35.466: INFO: Pod "downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5": Phase="Pending", Reason="", readiness=false. Elapsed: 5.625857ms -Dec 14 15:52:37.475: INFO: Pod "downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5": Phase="Running", Reason="", readiness=false. Elapsed: 2.0150151s -Dec 14 15:52:39.474: INFO: Pod "downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013590927s -STEP: Saw pod success 12/14/22 15:52:39.474 -Dec 14 15:52:39.474: INFO: Pod "downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5" satisfied condition "Succeeded or Failed" -Dec 14 15:52:39.479: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5 container client-container: -STEP: delete the pod 12/14/22 15:52:39.5 -Dec 14 15:52:39.518: INFO: Waiting for pod downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5 to disappear -Dec 14 15:52:39.523: INFO: Pod downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5 no longer exists -[AfterEach] [sig-storage] Projected downwardAPI +STEP: Creating a kubernetes client 02/25/23 09:52:39.403 +Feb 25 09:52:39.404: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 09:52:39.407 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:39.439 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:39.444 +[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:66 +STEP: Creating projection with secret that has name projected-secret-test-5cac0a69-e946-40e5-ad4a-87eb8439a109 02/25/23 09:52:39.452 +STEP: Creating a pod to test consume secrets 02/25/23 09:52:39.461 +Feb 25 09:52:39.474: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd" in namespace "projected-4783" to be "Succeeded or Failed" +Feb 25 09:52:39.482: INFO: Pod "pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd": Phase="Pending", Reason="", readiness=false. Elapsed: 7.725255ms +Feb 25 09:52:41.490: INFO: Pod "pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016016244s +Feb 25 09:52:43.492: INFO: Pod "pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018673349s +STEP: Saw pod success 02/25/23 09:52:43.493 +Feb 25 09:52:43.493: INFO: Pod "pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd" satisfied condition "Succeeded or Failed" +Feb 25 09:52:43.498: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd container projected-secret-volume-test: +STEP: delete the pod 02/25/23 09:52:43.51 +Feb 25 09:52:43.534: INFO: Waiting for pod pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd to disappear +Feb 25 09:52:43.540: INFO: Pod pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd no longer exists +[AfterEach] [sig-storage] Projected secret test/e2e/framework/framework.go:187 -Dec 14 15:52:39.523: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-2948" for this suite. 12/14/22 15:52:39.532 -{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance]","completed":30,"skipped":456,"failed":0} +Feb 25 09:52:43.540: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-4783" for this suite. 02/25/23 09:52:43.547 +{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","completed":31,"skipped":847,"failed":0} ------------------------------ -• [4.140 seconds] -[sig-storage] Projected downwardAPI +• [4.155 seconds] +[sig-storage] Projected secret test/e2e/common/storage/framework.go:23 - should provide container's memory request [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:234 + should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:66 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected downwardAPI + [BeforeEach] [sig-storage] Projected secret test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:52:35.399 - Dec 14 15:52:35.399: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 15:52:35.403 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:52:35.437 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:52:35.441 - [BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 - [It] should provide container's memory request [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:234 - STEP: Creating a pod to test downward API volume plugin 12/14/22 15:52:35.444 - Dec 14 15:52:35.460: INFO: Waiting up to 5m0s for pod "downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5" in namespace "projected-2948" to be "Succeeded or Failed" - Dec 14 15:52:35.466: INFO: Pod "downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5": Phase="Pending", Reason="", readiness=false. Elapsed: 5.625857ms - Dec 14 15:52:37.475: INFO: Pod "downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5": Phase="Running", Reason="", readiness=false. Elapsed: 2.0150151s - Dec 14 15:52:39.474: INFO: Pod "downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013590927s - STEP: Saw pod success 12/14/22 15:52:39.474 - Dec 14 15:52:39.474: INFO: Pod "downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5" satisfied condition "Succeeded or Failed" - Dec 14 15:52:39.479: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5 container client-container: - STEP: delete the pod 12/14/22 15:52:39.5 - Dec 14 15:52:39.518: INFO: Waiting for pod downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5 to disappear - Dec 14 15:52:39.523: INFO: Pod downwardapi-volume-8795bd66-0bca-4a54-91bd-fb7f50d759c5 no longer exists - [AfterEach] [sig-storage] Projected downwardAPI + STEP: Creating a kubernetes client 02/25/23 09:52:39.403 + Feb 25 09:52:39.404: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 09:52:39.407 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:39.439 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:39.444 + [It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:66 + STEP: Creating projection with secret that has name projected-secret-test-5cac0a69-e946-40e5-ad4a-87eb8439a109 02/25/23 09:52:39.452 + STEP: Creating a pod to test consume secrets 02/25/23 09:52:39.461 + Feb 25 09:52:39.474: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd" in namespace "projected-4783" to be "Succeeded or Failed" + Feb 25 09:52:39.482: INFO: Pod "pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd": Phase="Pending", Reason="", readiness=false. Elapsed: 7.725255ms + Feb 25 09:52:41.490: INFO: Pod "pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016016244s + Feb 25 09:52:43.492: INFO: Pod "pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018673349s + STEP: Saw pod success 02/25/23 09:52:43.493 + Feb 25 09:52:43.493: INFO: Pod "pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd" satisfied condition "Succeeded or Failed" + Feb 25 09:52:43.498: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd container projected-secret-volume-test: + STEP: delete the pod 02/25/23 09:52:43.51 + Feb 25 09:52:43.534: INFO: Waiting for pod pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd to disappear + Feb 25 09:52:43.540: INFO: Pod pod-projected-secrets-b842e9fd-3f3f-4ff5-abbf-22c3b0acdddd no longer exists + [AfterEach] [sig-storage] Projected secret test/e2e/framework/framework.go:187 - Dec 14 15:52:39.523: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-2948" for this suite. 12/14/22 15:52:39.532 + Feb 25 09:52:43.540: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-4783" for this suite. 02/25/23 09:52:43.547 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] - Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] - test/e2e/apps/statefulset.go:585 -[BeforeEach] [sig-apps] StatefulSet +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + works for multiple CRDs of same group but different versions [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:308 +[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:52:39.547 -Dec 14 15:52:39.548: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename statefulset 12/14/22 15:52:39.549 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:52:39.573 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:52:39.576 -[BeforeEach] [sig-apps] StatefulSet - test/e2e/apps/statefulset.go:96 -[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:111 -STEP: Creating service test in namespace statefulset-6163 12/14/22 15:52:39.579 -[It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] - test/e2e/apps/statefulset.go:585 -STEP: Initializing watcher for selector baz=blah,foo=bar 12/14/22 15:52:39.586 -STEP: Creating stateful set ss in namespace statefulset-6163 12/14/22 15:52:39.589 -STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-6163 12/14/22 15:52:39.596 -Dec 14 15:52:39.599: INFO: Found 0 stateful pods, waiting for 1 -Dec 14 15:52:49.612: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true -STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod 12/14/22 15:52:49.613 -Dec 14 15:52:49.618: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' -Dec 14 15:52:49.915: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" -Dec 14 15:52:49.915: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" -Dec 14 15:52:49.915: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - -Dec 14 15:52:49.923: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true -Dec 14 15:52:59.942: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false -Dec 14 15:52:59.943: INFO: Waiting for statefulset status.replicas updated to 0 -Dec 14 15:52:59.979: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999999616s -Dec 14 15:53:00.992: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.991403132s -Dec 14 15:53:01.999: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.977922969s -Dec 14 15:53:03.005: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.970614823s -Dec 14 15:53:04.012: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.965163788s -Dec 14 15:53:05.023: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.957975025s -Dec 14 15:53:06.033: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.947469218s -Dec 14 15:53:07.043: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.937214704s -Dec 14 15:53:08.052: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.927287957s -Dec 14 15:53:09.068: INFO: Verifying statefulset ss doesn't scale past 1 for another 918.734466ms -STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-6163 12/14/22 15:53:10.069 -Dec 14 15:53:10.079: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Dec 14 15:53:10.340: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" -Dec 14 15:53:10.341: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" -Dec 14 15:53:10.341: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - -Dec 14 15:53:10.346: INFO: Found 1 stateful pods, waiting for 3 -Dec 14 15:53:20.357: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true -Dec 14 15:53:20.357: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true -Dec 14 15:53:20.357: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true -STEP: Verifying that stateful set ss was scaled up in order 12/14/22 15:53:20.358 -STEP: Scale down will halt with unhealthy stateful pod 12/14/22 15:53:20.358 -Dec 14 15:53:20.370: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' -Dec 14 15:53:20.593: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" -Dec 14 15:53:20.593: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" -Dec 14 15:53:20.593: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - -Dec 14 15:53:20.593: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' -Dec 14 15:53:20.848: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" -Dec 14 15:53:20.849: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" -Dec 14 15:53:20.849: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - -Dec 14 15:53:20.850: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' -Dec 14 15:53:21.111: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" -Dec 14 15:53:21.111: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" -Dec 14 15:53:21.111: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - -Dec 14 15:53:21.111: INFO: Waiting for statefulset status.replicas updated to 0 -Dec 14 15:53:21.126: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 2 -Dec 14 15:53:31.144: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false -Dec 14 15:53:31.144: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false -Dec 14 15:53:31.144: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false -Dec 14 15:53:31.175: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999999567s -Dec 14 15:53:32.184: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.989098365s -Dec 14 15:53:33.194: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.980364775s -Dec 14 15:53:34.203: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.969780017s -Dec 14 15:53:35.214: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.960859929s -Dec 14 15:53:36.223: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.950259318s -Dec 14 15:53:37.235: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.941685347s -Dec 14 15:53:38.244: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.928716681s -Dec 14 15:53:39.253: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.919758652s -Dec 14 15:53:40.265: INFO: Verifying statefulset ss doesn't scale past 3 for another 910.708974ms -STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-6163 12/14/22 15:53:41.266 -Dec 14 15:53:41.278: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Dec 14 15:53:41.523: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" -Dec 14 15:53:41.523: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" -Dec 14 15:53:41.523: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - -Dec 14 15:53:41.523: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Dec 14 15:53:41.744: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" -Dec 14 15:53:41.744: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" -Dec 14 15:53:41.744: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - -Dec 14 15:53:41.745: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Dec 14 15:53:41.978: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" -Dec 14 15:53:41.978: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" -Dec 14 15:53:41.978: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - -Dec 14 15:53:41.978: INFO: Scaling statefulset ss to 0 -STEP: Verifying that stateful set ss was scaled down in reverse order 12/14/22 15:53:52.016 -[AfterEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:122 -Dec 14 15:53:52.018: INFO: Deleting all statefulset in ns statefulset-6163 -Dec 14 15:53:52.023: INFO: Scaling statefulset ss to 0 -Dec 14 15:53:52.040: INFO: Waiting for statefulset status.replicas updated to 0 -Dec 14 15:53:52.046: INFO: Deleting statefulset ss -[AfterEach] [sig-apps] StatefulSet +STEP: Creating a kubernetes client 02/25/23 09:52:43.561 +Feb 25 09:52:43.561: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 09:52:43.564 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:43.594 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:43.599 +[It] works for multiple CRDs of same group but different versions [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:308 +STEP: CRs in the same group but different versions (one multiversion CRD) show up in OpenAPI documentation 02/25/23 09:52:43.603 +Feb 25 09:52:43.604: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: CRs in the same group but different versions (two CRDs) show up in OpenAPI documentation 02/25/23 09:53:02.972 +Feb 25 09:53:02.975: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 09:53:06.849: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 15:53:52.075: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "statefulset-6163" for this suite. 12/14/22 15:53:52.082 -{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]","completed":31,"skipped":473,"failed":0} +Feb 25 09:53:27.118: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "crd-publish-openapi-6330" for this suite. 02/25/23 09:53:27.139 +{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group but different versions [Conformance]","completed":32,"skipped":884,"failed":0} ------------------------------ -• [SLOW TEST] [72.544 seconds] -[sig-apps] StatefulSet -test/e2e/apps/framework.go:23 - Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:101 - Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] - test/e2e/apps/statefulset.go:585 +• [SLOW TEST] [43.592 seconds] +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + works for multiple CRDs of same group but different versions [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:308 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] StatefulSet + [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:52:39.547 - Dec 14 15:52:39.548: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename statefulset 12/14/22 15:52:39.549 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:52:39.573 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:52:39.576 - [BeforeEach] [sig-apps] StatefulSet - test/e2e/apps/statefulset.go:96 - [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:111 - STEP: Creating service test in namespace statefulset-6163 12/14/22 15:52:39.579 - [It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] - test/e2e/apps/statefulset.go:585 - STEP: Initializing watcher for selector baz=blah,foo=bar 12/14/22 15:52:39.586 - STEP: Creating stateful set ss in namespace statefulset-6163 12/14/22 15:52:39.589 - STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-6163 12/14/22 15:52:39.596 - Dec 14 15:52:39.599: INFO: Found 0 stateful pods, waiting for 1 - Dec 14 15:52:49.612: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true - STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod 12/14/22 15:52:49.613 - Dec 14 15:52:49.618: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' - Dec 14 15:52:49.915: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" - Dec 14 15:52:49.915: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" - Dec 14 15:52:49.915: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - - Dec 14 15:52:49.923: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true - Dec 14 15:52:59.942: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false - Dec 14 15:52:59.943: INFO: Waiting for statefulset status.replicas updated to 0 - Dec 14 15:52:59.979: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999999616s - Dec 14 15:53:00.992: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.991403132s - Dec 14 15:53:01.999: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.977922969s - Dec 14 15:53:03.005: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.970614823s - Dec 14 15:53:04.012: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.965163788s - Dec 14 15:53:05.023: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.957975025s - Dec 14 15:53:06.033: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.947469218s - Dec 14 15:53:07.043: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.937214704s - Dec 14 15:53:08.052: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.927287957s - Dec 14 15:53:09.068: INFO: Verifying statefulset ss doesn't scale past 1 for another 918.734466ms - STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-6163 12/14/22 15:53:10.069 - Dec 14 15:53:10.079: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' - Dec 14 15:53:10.340: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" - Dec 14 15:53:10.341: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" - Dec 14 15:53:10.341: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - - Dec 14 15:53:10.346: INFO: Found 1 stateful pods, waiting for 3 - Dec 14 15:53:20.357: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true - Dec 14 15:53:20.357: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true - Dec 14 15:53:20.357: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true - STEP: Verifying that stateful set ss was scaled up in order 12/14/22 15:53:20.358 - STEP: Scale down will halt with unhealthy stateful pod 12/14/22 15:53:20.358 - Dec 14 15:53:20.370: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' - Dec 14 15:53:20.593: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" - Dec 14 15:53:20.593: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" - Dec 14 15:53:20.593: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - - Dec 14 15:53:20.593: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' - Dec 14 15:53:20.848: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" - Dec 14 15:53:20.849: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" - Dec 14 15:53:20.849: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - - Dec 14 15:53:20.850: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' - Dec 14 15:53:21.111: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" - Dec 14 15:53:21.111: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" - Dec 14 15:53:21.111: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - - Dec 14 15:53:21.111: INFO: Waiting for statefulset status.replicas updated to 0 - Dec 14 15:53:21.126: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 2 - Dec 14 15:53:31.144: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false - Dec 14 15:53:31.144: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false - Dec 14 15:53:31.144: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false - Dec 14 15:53:31.175: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999999567s - Dec 14 15:53:32.184: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.989098365s - Dec 14 15:53:33.194: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.980364775s - Dec 14 15:53:34.203: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.969780017s - Dec 14 15:53:35.214: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.960859929s - Dec 14 15:53:36.223: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.950259318s - Dec 14 15:53:37.235: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.941685347s - Dec 14 15:53:38.244: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.928716681s - Dec 14 15:53:39.253: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.919758652s - Dec 14 15:53:40.265: INFO: Verifying statefulset ss doesn't scale past 3 for another 910.708974ms - STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-6163 12/14/22 15:53:41.266 - Dec 14 15:53:41.278: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' - Dec 14 15:53:41.523: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" - Dec 14 15:53:41.523: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" - Dec 14 15:53:41.523: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - - Dec 14 15:53:41.523: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' - Dec 14 15:53:41.744: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" - Dec 14 15:53:41.744: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" - Dec 14 15:53:41.744: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - - Dec 14 15:53:41.745: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6163 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' - Dec 14 15:53:41.978: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" - Dec 14 15:53:41.978: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" - Dec 14 15:53:41.978: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - - Dec 14 15:53:41.978: INFO: Scaling statefulset ss to 0 - STEP: Verifying that stateful set ss was scaled down in reverse order 12/14/22 15:53:52.016 - [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:122 - Dec 14 15:53:52.018: INFO: Deleting all statefulset in ns statefulset-6163 - Dec 14 15:53:52.023: INFO: Scaling statefulset ss to 0 - Dec 14 15:53:52.040: INFO: Waiting for statefulset status.replicas updated to 0 - Dec 14 15:53:52.046: INFO: Deleting statefulset ss - [AfterEach] [sig-apps] StatefulSet + STEP: Creating a kubernetes client 02/25/23 09:52:43.561 + Feb 25 09:52:43.561: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 09:52:43.564 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:52:43.594 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:52:43.599 + [It] works for multiple CRDs of same group but different versions [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:308 + STEP: CRs in the same group but different versions (one multiversion CRD) show up in OpenAPI documentation 02/25/23 09:52:43.603 + Feb 25 09:52:43.604: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: CRs in the same group but different versions (two CRDs) show up in OpenAPI documentation 02/25/23 09:53:02.972 + Feb 25 09:53:02.975: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 09:53:06.849: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 15:53:52.075: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "statefulset-6163" for this suite. 12/14/22 15:53:52.082 + Feb 25 09:53:27.118: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "crd-publish-openapi-6330" for this suite. 02/25/23 09:53:27.139 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSS +SSSSSSSSS ------------------------------ -[sig-network] DNS - should provide DNS for pods for Hostname [Conformance] - test/e2e/network/dns.go:248 -[BeforeEach] [sig-network] DNS +[sig-storage] ConfigMap + should be immutable if `immutable` field is set [Conformance] + test/e2e/common/storage/configmap_volume.go:503 +[BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:53:52.1 -Dec 14 15:53:52.100: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename dns 12/14/22 15:53:52.104 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:53:52.124 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:53:52.129 -[It] should provide DNS for pods for Hostname [Conformance] - test/e2e/network/dns.go:248 -STEP: Creating a test headless service 12/14/22 15:53:52.134 -STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-8322.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-2.dns-test-service-2.dns-8322.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/wheezy_hosts@dns-querier-2;sleep 1; done - 12/14/22 15:53:52.143 -STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-8322.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-2.dns-test-service-2.dns-8322.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/jessie_hosts@dns-querier-2;sleep 1; done - 12/14/22 15:53:52.143 -STEP: creating a pod to probe DNS 12/14/22 15:53:52.143 -STEP: submitting the pod to kubernetes 12/14/22 15:53:52.144 -Dec 14 15:53:52.161: INFO: Waiting up to 15m0s for pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781" in namespace "dns-8322" to be "running" -Dec 14 15:53:52.166: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Pending", Reason="", readiness=false. Elapsed: 4.940877ms -Dec 14 15:53:54.176: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014958105s -Dec 14 15:53:56.188: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Pending", Reason="", readiness=false. Elapsed: 4.027533703s -Dec 14 15:53:58.172: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Pending", Reason="", readiness=false. Elapsed: 6.01161359s -Dec 14 15:54:00.173: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Pending", Reason="", readiness=false. Elapsed: 8.011843141s -Dec 14 15:54:02.173: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Pending", Reason="", readiness=false. Elapsed: 10.01238958s -Dec 14 15:54:04.174: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Pending", Reason="", readiness=false. Elapsed: 12.013205512s -Dec 14 15:54:06.175: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Running", Reason="", readiness=true. Elapsed: 14.014738972s -Dec 14 15:54:06.176: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781" satisfied condition "running" -STEP: retrieving the pod 12/14/22 15:54:06.176 -STEP: looking for the results for each expected name from probers 12/14/22 15:54:06.182 -Dec 14 15:54:06.219: INFO: DNS probes using dns-8322/dns-test-dff79909-b009-4ba9-a59f-7b302e935781 succeeded - -STEP: deleting the pod 12/14/22 15:54:06.219 -STEP: deleting the test headless service 12/14/22 15:54:06.244 -[AfterEach] [sig-network] DNS +STEP: Creating a kubernetes client 02/25/23 09:53:27.156 +Feb 25 09:53:27.156: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename configmap 02/25/23 09:53:27.161 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:53:27.187 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:53:27.193 +[It] should be immutable if `immutable` field is set [Conformance] + test/e2e/common/storage/configmap_volume.go:503 +[AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 -Dec 14 15:54:06.262: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "dns-8322" for this suite. 12/14/22 15:54:06.268 -{"msg":"PASSED [sig-network] DNS should provide DNS for pods for Hostname [Conformance]","completed":32,"skipped":483,"failed":0} +Feb 25 09:53:27.259: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-2860" for this suite. 02/25/23 09:53:27.266 +{"msg":"PASSED [sig-storage] ConfigMap should be immutable if `immutable` field is set [Conformance]","completed":33,"skipped":893,"failed":0} ------------------------------ -• [SLOW TEST] [14.178 seconds] -[sig-network] DNS -test/e2e/network/common/framework.go:23 - should provide DNS for pods for Hostname [Conformance] - test/e2e/network/dns.go:248 +• [0.123 seconds] +[sig-storage] ConfigMap +test/e2e/common/storage/framework.go:23 + should be immutable if `immutable` field is set [Conformance] + test/e2e/common/storage/configmap_volume.go:503 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] DNS + [BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:53:52.1 - Dec 14 15:53:52.100: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename dns 12/14/22 15:53:52.104 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:53:52.124 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:53:52.129 - [It] should provide DNS for pods for Hostname [Conformance] - test/e2e/network/dns.go:248 - STEP: Creating a test headless service 12/14/22 15:53:52.134 - STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-8322.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-2.dns-test-service-2.dns-8322.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/wheezy_hosts@dns-querier-2;sleep 1; done - 12/14/22 15:53:52.143 - STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-8322.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-2.dns-test-service-2.dns-8322.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/jessie_hosts@dns-querier-2;sleep 1; done - 12/14/22 15:53:52.143 - STEP: creating a pod to probe DNS 12/14/22 15:53:52.143 - STEP: submitting the pod to kubernetes 12/14/22 15:53:52.144 - Dec 14 15:53:52.161: INFO: Waiting up to 15m0s for pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781" in namespace "dns-8322" to be "running" - Dec 14 15:53:52.166: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Pending", Reason="", readiness=false. Elapsed: 4.940877ms - Dec 14 15:53:54.176: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014958105s - Dec 14 15:53:56.188: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Pending", Reason="", readiness=false. Elapsed: 4.027533703s - Dec 14 15:53:58.172: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Pending", Reason="", readiness=false. Elapsed: 6.01161359s - Dec 14 15:54:00.173: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Pending", Reason="", readiness=false. Elapsed: 8.011843141s - Dec 14 15:54:02.173: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Pending", Reason="", readiness=false. Elapsed: 10.01238958s - Dec 14 15:54:04.174: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Pending", Reason="", readiness=false. Elapsed: 12.013205512s - Dec 14 15:54:06.175: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781": Phase="Running", Reason="", readiness=true. Elapsed: 14.014738972s - Dec 14 15:54:06.176: INFO: Pod "dns-test-dff79909-b009-4ba9-a59f-7b302e935781" satisfied condition "running" - STEP: retrieving the pod 12/14/22 15:54:06.176 - STEP: looking for the results for each expected name from probers 12/14/22 15:54:06.182 - Dec 14 15:54:06.219: INFO: DNS probes using dns-8322/dns-test-dff79909-b009-4ba9-a59f-7b302e935781 succeeded - - STEP: deleting the pod 12/14/22 15:54:06.219 - STEP: deleting the test headless service 12/14/22 15:54:06.244 - [AfterEach] [sig-network] DNS + STEP: Creating a kubernetes client 02/25/23 09:53:27.156 + Feb 25 09:53:27.156: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename configmap 02/25/23 09:53:27.161 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:53:27.187 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:53:27.193 + [It] should be immutable if `immutable` field is set [Conformance] + test/e2e/common/storage/configmap_volume.go:503 + [AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 - Dec 14 15:54:06.262: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "dns-8322" for this suite. 12/14/22 15:54:06.268 + Feb 25 09:53:27.259: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "configmap-2860" for this suite. 02/25/23 09:53:27.266 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSS +SS ------------------------------ -[sig-apps] Deployment - should run the lifecycle of a Deployment [Conformance] - test/e2e/apps/deployment.go:185 -[BeforeEach] [sig-apps] Deployment +[sig-node] Variable Expansion + should verify that a failing subpath expansion can be modified during the lifecycle of a container [Slow] [Conformance] + test/e2e/common/node/expansion.go:224 +[BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:54:06.279 -Dec 14 15:54:06.279: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename deployment 12/14/22 15:54:06.282 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:54:06.308 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:54:06.312 -[BeforeEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:91 -[It] should run the lifecycle of a Deployment [Conformance] - test/e2e/apps/deployment.go:185 -STEP: creating a Deployment 12/14/22 15:54:06.321 -STEP: waiting for Deployment to be created 12/14/22 15:54:06.33 -STEP: waiting for all Replicas to be Ready 12/14/22 15:54:06.332 -Dec 14 15:54:06.334: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] -Dec 14 15:54:06.335: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] -Dec 14 15:54:06.349: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] -Dec 14 15:54:06.350: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] -Dec 14 15:54:06.377: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] -Dec 14 15:54:06.378: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] -Dec 14 15:54:06.440: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] -Dec 14 15:54:06.440: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] -Dec 14 15:54:07.270: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment-static:true] -Dec 14 15:54:07.270: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment-static:true] -Dec 14 15:54:07.976: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 and labels map[test-deployment-static:true] -STEP: patching the Deployment 12/14/22 15:54:07.976 -W1214 15:54:07.993015 14 warnings.go:70] unknown field "spec.template.spec.TerminationGracePeriodSeconds" -Dec 14 15:54:07.996: INFO: observed event type ADDED -STEP: waiting for Replicas to scale 12/14/22 15:54:07.996 -Dec 14 15:54:08.000: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 -Dec 14 15:54:08.000: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 -Dec 14 15:54:08.000: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 -Dec 14 15:54:08.001: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 -Dec 14 15:54:08.001: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 -Dec 14 15:54:08.001: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 -Dec 14 15:54:08.001: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 -Dec 14 15:54:08.001: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 -Dec 14 15:54:08.001: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 -Dec 14 15:54:08.002: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 -Dec 14 15:54:08.002: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 -Dec 14 15:54:08.002: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 -Dec 14 15:54:08.002: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 -Dec 14 15:54:08.002: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 -Dec 14 15:54:08.012: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 -Dec 14 15:54:08.012: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 -Dec 14 15:54:08.040: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 -Dec 14 15:54:08.040: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 -Dec 14 15:54:08.071: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 -Dec 14 15:54:08.071: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 -Dec 14 15:54:08.080: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 -Dec 14 15:54:08.080: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 -Dec 14 15:54:08.987: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 -Dec 14 15:54:08.987: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 -Dec 14 15:54:09.037: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 -STEP: listing Deployments 12/14/22 15:54:09.037 -Dec 14 15:54:09.044: INFO: Found test-deployment with labels: map[test-deployment:patched test-deployment-static:true] -STEP: updating the Deployment 12/14/22 15:54:09.044 -Dec 14 15:54:09.071: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 -STEP: fetching the DeploymentStatus 12/14/22 15:54:09.071 -Dec 14 15:54:09.080: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] -Dec 14 15:54:09.104: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] -Dec 14 15:54:09.129: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] -Dec 14 15:54:09.179: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] -Dec 14 15:54:09.197: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] -Dec 14 15:54:09.219: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] -Dec 14 15:54:09.845: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] -Dec 14 15:54:10.017: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 3 and labels map[test-deployment:updated test-deployment-static:true] -Dec 14 15:54:10.107: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] -Dec 14 15:54:10.133: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] -Dec 14 15:54:11.280: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 3 and labels map[test-deployment:updated test-deployment-static:true] -STEP: patching the DeploymentStatus 12/14/22 15:54:11.337 -STEP: fetching the DeploymentStatus 12/14/22 15:54:11.352 -Dec 14 15:54:11.361: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 -Dec 14 15:54:11.361: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 -Dec 14 15:54:11.362: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 -Dec 14 15:54:11.362: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 -Dec 14 15:54:11.362: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 -Dec 14 15:54:11.362: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 -Dec 14 15:54:11.363: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 -Dec 14 15:54:11.363: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 3 -Dec 14 15:54:11.363: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 -Dec 14 15:54:11.363: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 -Dec 14 15:54:11.363: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 3 -STEP: deleting the Deployment 12/14/22 15:54:11.364 -Dec 14 15:54:11.377: INFO: observed event type MODIFIED -Dec 14 15:54:11.377: INFO: observed event type MODIFIED -Dec 14 15:54:11.377: INFO: observed event type MODIFIED -Dec 14 15:54:11.378: INFO: observed event type MODIFIED -Dec 14 15:54:11.378: INFO: observed event type MODIFIED -Dec 14 15:54:11.378: INFO: observed event type MODIFIED -Dec 14 15:54:11.379: INFO: observed event type MODIFIED -Dec 14 15:54:11.379: INFO: observed event type MODIFIED -Dec 14 15:54:11.379: INFO: observed event type MODIFIED -Dec 14 15:54:11.379: INFO: observed event type MODIFIED -Dec 14 15:54:11.380: INFO: observed event type MODIFIED -Dec 14 15:54:11.380: INFO: observed event type MODIFIED -[AfterEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:84 -Dec 14 15:54:11.386: INFO: Log out all the ReplicaSets if there is no deployment created -Dec 14 15:54:11.391: INFO: ReplicaSet "test-deployment-54cc775c4b": -&ReplicaSet{ObjectMeta:{test-deployment-54cc775c4b deployment-1216 05f2db87-7fb5-4b52-9ad6-8b0479ce9b2e 6154 4 2022-12-14 15:54:07 +0000 UTC map[pod-template-hash:54cc775c4b test-deployment-static:true] map[deployment.kubernetes.io/desired-replicas:2 deployment.kubernetes.io/max-replicas:3 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-deployment 5fec78d8-938b-4640-bd31-571abea9033e 0xc00216a157 0xc00216a158}] [] [{kube-controller-manager Update apps/v1 2022-12-14 15:54:11 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"5fec78d8-938b-4640-bd31-571abea9033e\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 15:54:11 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{pod-template-hash: 54cc775c4b,test-deployment-static: true,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[pod-template-hash:54cc775c4b test-deployment-static:true] map[] [] [] []} {[] [] [{test-deployment registry.k8s.io/pause:3.8 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc00216a1e0 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:4,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} - -Dec 14 15:54:11.401: INFO: pod: "test-deployment-54cc775c4b-2tgf7": -&Pod{ObjectMeta:{test-deployment-54cc775c4b-2tgf7 test-deployment-54cc775c4b- deployment-1216 af0e43ba-89e7-499e-b80d-d318166b2aca 6116 0 2022-12-14 15:54:08 +0000 UTC 2022-12-14 15:54:10 +0000 UTC 0xc00216a640 map[pod-template-hash:54cc775c4b test-deployment-static:true] map[] [{apps/v1 ReplicaSet test-deployment-54cc775c4b 05f2db87-7fb5-4b52-9ad6-8b0479ce9b2e 0xc00216a677 0xc00216a678}] [] [{kube-controller-manager Update v1 2022-12-14 15:54:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"05f2db87-7fb5-4b52-9ad6-8b0479ce9b2e\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 15:54:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.44\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-glclm,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:test-deployment,Image:registry.k8s.io/pause:3.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-glclm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*1,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.44,StartTime:2022-12-14 15:54:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:test-deployment,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 15:54:08 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/pause:3.8,ImageID:registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d,ContainerID:cri-o://a7b10b14acd27564f3442f4d2e59d254431ccf14ca83a1d5f949f625c46eb533,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.44,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - -Dec 14 15:54:11.401: INFO: pod: "test-deployment-54cc775c4b-hpqc5": -&Pod{ObjectMeta:{test-deployment-54cc775c4b-hpqc5 test-deployment-54cc775c4b- deployment-1216 a360eb18-7c8b-4995-a3af-7d6322ff75c6 6151 0 2022-12-14 15:54:09 +0000 UTC 2022-12-14 15:54:12 +0000 UTC 0xc00216a840 map[pod-template-hash:54cc775c4b test-deployment-static:true] map[] [{apps/v1 ReplicaSet test-deployment-54cc775c4b 05f2db87-7fb5-4b52-9ad6-8b0479ce9b2e 0xc00216a877 0xc00216a878}] [] [{kube-controller-manager Update v1 2022-12-14 15:54:09 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"05f2db87-7fb5-4b52-9ad6-8b0479ce9b2e\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 15:54:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.16\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-5d2m9,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:test-deployment,Image:registry.k8s.io/pause:3.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5d2m9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*1,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:10.233.65.16,StartTime:2022-12-14 15:54:09 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:test-deployment,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 15:54:09 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/pause:3.8,ImageID:registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d,ContainerID:cri-o://652d03376ff2e8a767f128061d71285881f209ea3207d83694c2e4b565960e24,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.16,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - -Dec 14 15:54:11.402: INFO: ReplicaSet "test-deployment-7c7d8d58c8": -&ReplicaSet{ObjectMeta:{test-deployment-7c7d8d58c8 deployment-1216 05308084-3ea1-4062-9c29-bd6071e17afe 6147 2 2022-12-14 15:54:09 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[deployment.kubernetes.io/desired-replicas:2 deployment.kubernetes.io/max-replicas:3 deployment.kubernetes.io/revision:3] [{apps/v1 Deployment test-deployment 5fec78d8-938b-4640-bd31-571abea9033e 0xc00216a247 0xc00216a248}] [] [{kube-controller-manager Update apps/v1 2022-12-14 15:54:10 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"5fec78d8-938b-4640-bd31-571abea9033e\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 15:54:11 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*2,Selector:&v1.LabelSelector{MatchLabels:map[string]string{pod-template-hash: 7c7d8d58c8,test-deployment-static: true,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[] [] [] []} {[] [] [{test-deployment registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc00216a2e0 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:2,FullyLabeledReplicas:2,ObservedGeneration:2,ReadyReplicas:2,AvailableReplicas:2,Conditions:[]ReplicaSetCondition{},},} - -Dec 14 15:54:11.407: INFO: pod: "test-deployment-7c7d8d58c8-5fzd2": -&Pod{ObjectMeta:{test-deployment-7c7d8d58c8-5fzd2 test-deployment-7c7d8d58c8- deployment-1216 6fde8472-2be0-4711-8138-125cc950cbdb 6111 0 2022-12-14 15:54:09 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[] [{apps/v1 ReplicaSet test-deployment-7c7d8d58c8 05308084-3ea1-4062-9c29-bd6071e17afe 0xc002920a67 0xc002920a68}] [] [{kube-controller-manager Update v1 2022-12-14 15:54:09 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"05308084-3ea1-4062-9c29-bd6071e17afe\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 15:54:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.45\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-8tn6q,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:test-deployment,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8tn6q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*1,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.45,StartTime:2022-12-14 15:54:09 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:test-deployment,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 15:54:09 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://22f7a860436c766ba8a7c3133d35ada2acee6d7a15455a317366ca969516ce2e,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.45,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - -Dec 14 15:54:11.407: INFO: pod: "test-deployment-7c7d8d58c8-6kskv": -&Pod{ObjectMeta:{test-deployment-7c7d8d58c8-6kskv test-deployment-7c7d8d58c8- deployment-1216 d16be486-e0da-4f3c-86ac-291bfad58659 6146 0 2022-12-14 15:54:10 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[] [{apps/v1 ReplicaSet test-deployment-7c7d8d58c8 05308084-3ea1-4062-9c29-bd6071e17afe 0xc002920c67 0xc002920c68}] [] [{kube-controller-manager Update v1 2022-12-14 15:54:10 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"05308084-3ea1-4062-9c29-bd6071e17afe\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 15:54:11 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.13\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-2cltv,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:test-deployment,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2cltv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*1,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:11 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:11 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:10 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:10.233.64.13,StartTime:2022-12-14 15:54:10 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:test-deployment,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 15:54:10 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://8af110942d3d090b0ece07a95f0e9ab6e3e8d3d0c93508ace700d146d9d886b5,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.13,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - -Dec 14 15:54:11.408: INFO: ReplicaSet "test-deployment-8594bb6fdd": -&ReplicaSet{ObjectMeta:{test-deployment-8594bb6fdd deployment-1216 afb408ea-ad0a-4318-8204-af21e8672f0f 6074 3 2022-12-14 15:54:06 +0000 UTC map[pod-template-hash:8594bb6fdd test-deployment-static:true] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-deployment 5fec78d8-938b-4640-bd31-571abea9033e 0xc00216a347 0xc00216a348}] [] [{kube-controller-manager Update apps/v1 2022-12-14 15:54:08 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"5fec78d8-938b-4640-bd31-571abea9033e\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 15:54:08 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{pod-template-hash: 8594bb6fdd,test-deployment-static: true,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[pod-template-hash:8594bb6fdd test-deployment-static:true] map[] [] [] []} {[] [] [{test-deployment registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc00216a3d0 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} - -[AfterEach] [sig-apps] Deployment +STEP: Creating a kubernetes client 02/25/23 09:53:27.281 +Feb 25 09:53:27.281: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename var-expansion 02/25/23 09:53:27.284 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:53:27.311 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:53:27.315 +[It] should verify that a failing subpath expansion can be modified during the lifecycle of a container [Slow] [Conformance] + test/e2e/common/node/expansion.go:224 +STEP: creating the pod with failed condition 02/25/23 09:53:27.327 +Feb 25 09:53:27.346: INFO: Waiting up to 2m0s for pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92" in namespace "var-expansion-3821" to be "running" +Feb 25 09:53:27.353: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 7.395022ms +Feb 25 09:53:29.366: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019516304s +Feb 25 09:53:31.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 4.015993588s +Feb 25 09:53:33.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 6.016088713s +Feb 25 09:53:35.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 8.01645111s +Feb 25 09:53:37.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 10.014524243s +Feb 25 09:53:39.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 12.013531144s +Feb 25 09:53:41.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 14.015196458s +Feb 25 09:53:43.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 16.013769153s +Feb 25 09:53:45.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 18.014965925s +Feb 25 09:53:47.359: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 20.012613608s +Feb 25 09:53:49.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 22.017262996s +Feb 25 09:53:51.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 24.017538904s +Feb 25 09:53:53.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 26.015213962s +Feb 25 09:53:55.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 28.018218849s +Feb 25 09:53:57.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 30.018031713s +Feb 25 09:53:59.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 32.014731889s +Feb 25 09:54:01.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 34.017081872s +Feb 25 09:54:03.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 36.015902906s +Feb 25 09:54:05.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 38.013930985s +Feb 25 09:54:07.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 40.014905028s +Feb 25 09:54:09.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 42.015621935s +Feb 25 09:54:11.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 44.016516021s +Feb 25 09:54:13.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 46.017632401s +Feb 25 09:54:15.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 48.015988989s +Feb 25 09:54:17.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 50.016509165s +Feb 25 09:54:19.365: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 52.019334567s +Feb 25 09:54:21.365: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 54.019274123s +Feb 25 09:54:23.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 56.015533614s +Feb 25 09:54:25.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 58.017517181s +Feb 25 09:54:27.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m0.016979961s +Feb 25 09:54:29.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m2.017387252s +Feb 25 09:54:31.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m4.014784211s +Feb 25 09:54:33.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m6.016313129s +Feb 25 09:54:35.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m8.015697455s +Feb 25 09:54:37.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m10.016457409s +Feb 25 09:54:39.359: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m12.012987704s +Feb 25 09:54:41.359: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m14.013363148s +Feb 25 09:54:43.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m16.014454441s +Feb 25 09:54:45.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m18.0181947s +Feb 25 09:54:47.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m20.01418567s +Feb 25 09:54:49.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m22.01502728s +Feb 25 09:54:51.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m24.018020774s +Feb 25 09:54:53.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m26.01391261s +Feb 25 09:54:55.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m28.015461151s +Feb 25 09:54:57.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m30.013652303s +Feb 25 09:54:59.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m32.016909507s +Feb 25 09:55:01.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m34.018446007s +Feb 25 09:55:03.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m36.015210318s +Feb 25 09:55:05.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m38.015723197s +Feb 25 09:55:07.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m40.015123919s +Feb 25 09:55:09.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m42.014854669s +Feb 25 09:55:11.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m44.015474257s +Feb 25 09:55:13.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m46.015208531s +Feb 25 09:55:15.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m48.015788742s +Feb 25 09:55:17.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m50.013513734s +Feb 25 09:55:19.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m52.013515429s +Feb 25 09:55:21.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m54.015952129s +Feb 25 09:55:23.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m56.014315363s +Feb 25 09:55:25.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m58.0149816s +Feb 25 09:55:27.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 2m0.017308997s +Feb 25 09:55:27.372: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 2m0.026252215s +STEP: updating the pod 02/25/23 09:55:27.373 +Feb 25 09:55:27.898: INFO: Successfully updated pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92" +STEP: waiting for pod running 02/25/23 09:55:27.898 +Feb 25 09:55:27.899: INFO: Waiting up to 2m0s for pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92" in namespace "var-expansion-3821" to be "running" +Feb 25 09:55:27.908: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 8.936321ms +Feb 25 09:55:29.917: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Running", Reason="", readiness=true. Elapsed: 2.018493087s +Feb 25 09:55:29.917: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92" satisfied condition "running" +STEP: deleting the pod gracefully 02/25/23 09:55:29.917 +Feb 25 09:55:29.918: INFO: Deleting pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92" in namespace "var-expansion-3821" +Feb 25 09:55:29.935: INFO: Wait up to 5m0s for pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92" to be fully deleted +[AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 -Dec 14 15:54:11.414: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "deployment-1216" for this suite. 12/14/22 15:54:11.423 -{"msg":"PASSED [sig-apps] Deployment should run the lifecycle of a Deployment [Conformance]","completed":33,"skipped":492,"failed":0} +Feb 25 09:56:01.953: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "var-expansion-3821" for this suite. 02/25/23 09:56:01.963 +{"msg":"PASSED [sig-node] Variable Expansion should verify that a failing subpath expansion can be modified during the lifecycle of a container [Slow] [Conformance]","completed":34,"skipped":895,"failed":0} ------------------------------ -• [SLOW TEST] [5.153 seconds] -[sig-apps] Deployment -test/e2e/apps/framework.go:23 - should run the lifecycle of a Deployment [Conformance] - test/e2e/apps/deployment.go:185 +• [SLOW TEST] [154.692 seconds] +[sig-node] Variable Expansion +test/e2e/common/node/framework.go:23 + should verify that a failing subpath expansion can be modified during the lifecycle of a container [Slow] [Conformance] + test/e2e/common/node/expansion.go:224 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Deployment + [BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:54:06.279 - Dec 14 15:54:06.279: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename deployment 12/14/22 15:54:06.282 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:54:06.308 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:54:06.312 - [BeforeEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:91 - [It] should run the lifecycle of a Deployment [Conformance] - test/e2e/apps/deployment.go:185 - STEP: creating a Deployment 12/14/22 15:54:06.321 - STEP: waiting for Deployment to be created 12/14/22 15:54:06.33 - STEP: waiting for all Replicas to be Ready 12/14/22 15:54:06.332 - Dec 14 15:54:06.334: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] - Dec 14 15:54:06.335: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] - Dec 14 15:54:06.349: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] - Dec 14 15:54:06.350: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] - Dec 14 15:54:06.377: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] - Dec 14 15:54:06.378: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] - Dec 14 15:54:06.440: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] - Dec 14 15:54:06.440: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 and labels map[test-deployment-static:true] - Dec 14 15:54:07.270: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment-static:true] - Dec 14 15:54:07.270: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment-static:true] - Dec 14 15:54:07.976: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 and labels map[test-deployment-static:true] - STEP: patching the Deployment 12/14/22 15:54:07.976 - W1214 15:54:07.993015 14 warnings.go:70] unknown field "spec.template.spec.TerminationGracePeriodSeconds" - Dec 14 15:54:07.996: INFO: observed event type ADDED - STEP: waiting for Replicas to scale 12/14/22 15:54:07.996 - Dec 14 15:54:08.000: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 - Dec 14 15:54:08.000: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 - Dec 14 15:54:08.000: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 - Dec 14 15:54:08.001: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 - Dec 14 15:54:08.001: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 - Dec 14 15:54:08.001: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 - Dec 14 15:54:08.001: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 - Dec 14 15:54:08.001: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 0 - Dec 14 15:54:08.001: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 - Dec 14 15:54:08.002: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 - Dec 14 15:54:08.002: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 - Dec 14 15:54:08.002: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 - Dec 14 15:54:08.002: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 - Dec 14 15:54:08.002: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 - Dec 14 15:54:08.012: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 - Dec 14 15:54:08.012: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 - Dec 14 15:54:08.040: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 - Dec 14 15:54:08.040: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 - Dec 14 15:54:08.071: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 - Dec 14 15:54:08.071: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 - Dec 14 15:54:08.080: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 - Dec 14 15:54:08.080: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 - Dec 14 15:54:08.987: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 - Dec 14 15:54:08.987: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 - Dec 14 15:54:09.037: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 - STEP: listing Deployments 12/14/22 15:54:09.037 - Dec 14 15:54:09.044: INFO: Found test-deployment with labels: map[test-deployment:patched test-deployment-static:true] - STEP: updating the Deployment 12/14/22 15:54:09.044 - Dec 14 15:54:09.071: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 - STEP: fetching the DeploymentStatus 12/14/22 15:54:09.071 - Dec 14 15:54:09.080: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] - Dec 14 15:54:09.104: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] - Dec 14 15:54:09.129: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] - Dec 14 15:54:09.179: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] - Dec 14 15:54:09.197: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] - Dec 14 15:54:09.219: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] - Dec 14 15:54:09.845: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] - Dec 14 15:54:10.017: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 3 and labels map[test-deployment:updated test-deployment-static:true] - Dec 14 15:54:10.107: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] - Dec 14 15:54:10.133: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] - Dec 14 15:54:11.280: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 3 and labels map[test-deployment:updated test-deployment-static:true] - STEP: patching the DeploymentStatus 12/14/22 15:54:11.337 - STEP: fetching the DeploymentStatus 12/14/22 15:54:11.352 - Dec 14 15:54:11.361: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 - Dec 14 15:54:11.361: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 - Dec 14 15:54:11.362: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 - Dec 14 15:54:11.362: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 - Dec 14 15:54:11.362: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 - Dec 14 15:54:11.362: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 1 - Dec 14 15:54:11.363: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 - Dec 14 15:54:11.363: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 3 - Dec 14 15:54:11.363: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 - Dec 14 15:54:11.363: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 2 - Dec 14 15:54:11.363: INFO: observed Deployment test-deployment in namespace deployment-1216 with ReadyReplicas 3 - STEP: deleting the Deployment 12/14/22 15:54:11.364 - Dec 14 15:54:11.377: INFO: observed event type MODIFIED - Dec 14 15:54:11.377: INFO: observed event type MODIFIED - Dec 14 15:54:11.377: INFO: observed event type MODIFIED - Dec 14 15:54:11.378: INFO: observed event type MODIFIED - Dec 14 15:54:11.378: INFO: observed event type MODIFIED - Dec 14 15:54:11.378: INFO: observed event type MODIFIED - Dec 14 15:54:11.379: INFO: observed event type MODIFIED - Dec 14 15:54:11.379: INFO: observed event type MODIFIED - Dec 14 15:54:11.379: INFO: observed event type MODIFIED - Dec 14 15:54:11.379: INFO: observed event type MODIFIED - Dec 14 15:54:11.380: INFO: observed event type MODIFIED - Dec 14 15:54:11.380: INFO: observed event type MODIFIED - [AfterEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:84 - Dec 14 15:54:11.386: INFO: Log out all the ReplicaSets if there is no deployment created - Dec 14 15:54:11.391: INFO: ReplicaSet "test-deployment-54cc775c4b": - &ReplicaSet{ObjectMeta:{test-deployment-54cc775c4b deployment-1216 05f2db87-7fb5-4b52-9ad6-8b0479ce9b2e 6154 4 2022-12-14 15:54:07 +0000 UTC map[pod-template-hash:54cc775c4b test-deployment-static:true] map[deployment.kubernetes.io/desired-replicas:2 deployment.kubernetes.io/max-replicas:3 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-deployment 5fec78d8-938b-4640-bd31-571abea9033e 0xc00216a157 0xc00216a158}] [] [{kube-controller-manager Update apps/v1 2022-12-14 15:54:11 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"5fec78d8-938b-4640-bd31-571abea9033e\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 15:54:11 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{pod-template-hash: 54cc775c4b,test-deployment-static: true,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[pod-template-hash:54cc775c4b test-deployment-static:true] map[] [] [] []} {[] [] [{test-deployment registry.k8s.io/pause:3.8 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc00216a1e0 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:4,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} - - Dec 14 15:54:11.401: INFO: pod: "test-deployment-54cc775c4b-2tgf7": - &Pod{ObjectMeta:{test-deployment-54cc775c4b-2tgf7 test-deployment-54cc775c4b- deployment-1216 af0e43ba-89e7-499e-b80d-d318166b2aca 6116 0 2022-12-14 15:54:08 +0000 UTC 2022-12-14 15:54:10 +0000 UTC 0xc00216a640 map[pod-template-hash:54cc775c4b test-deployment-static:true] map[] [{apps/v1 ReplicaSet test-deployment-54cc775c4b 05f2db87-7fb5-4b52-9ad6-8b0479ce9b2e 0xc00216a677 0xc00216a678}] [] [{kube-controller-manager Update v1 2022-12-14 15:54:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"05f2db87-7fb5-4b52-9ad6-8b0479ce9b2e\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 15:54:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.44\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-glclm,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:test-deployment,Image:registry.k8s.io/pause:3.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-glclm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*1,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.44,StartTime:2022-12-14 15:54:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:test-deployment,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 15:54:08 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/pause:3.8,ImageID:registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d,ContainerID:cri-o://a7b10b14acd27564f3442f4d2e59d254431ccf14ca83a1d5f949f625c46eb533,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.44,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - - Dec 14 15:54:11.401: INFO: pod: "test-deployment-54cc775c4b-hpqc5": - &Pod{ObjectMeta:{test-deployment-54cc775c4b-hpqc5 test-deployment-54cc775c4b- deployment-1216 a360eb18-7c8b-4995-a3af-7d6322ff75c6 6151 0 2022-12-14 15:54:09 +0000 UTC 2022-12-14 15:54:12 +0000 UTC 0xc00216a840 map[pod-template-hash:54cc775c4b test-deployment-static:true] map[] [{apps/v1 ReplicaSet test-deployment-54cc775c4b 05f2db87-7fb5-4b52-9ad6-8b0479ce9b2e 0xc00216a877 0xc00216a878}] [] [{kube-controller-manager Update v1 2022-12-14 15:54:09 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"05f2db87-7fb5-4b52-9ad6-8b0479ce9b2e\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 15:54:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.16\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-5d2m9,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:test-deployment,Image:registry.k8s.io/pause:3.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5d2m9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*1,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:10.233.65.16,StartTime:2022-12-14 15:54:09 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:test-deployment,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 15:54:09 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/pause:3.8,ImageID:registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d,ContainerID:cri-o://652d03376ff2e8a767f128061d71285881f209ea3207d83694c2e4b565960e24,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.16,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - - Dec 14 15:54:11.402: INFO: ReplicaSet "test-deployment-7c7d8d58c8": - &ReplicaSet{ObjectMeta:{test-deployment-7c7d8d58c8 deployment-1216 05308084-3ea1-4062-9c29-bd6071e17afe 6147 2 2022-12-14 15:54:09 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[deployment.kubernetes.io/desired-replicas:2 deployment.kubernetes.io/max-replicas:3 deployment.kubernetes.io/revision:3] [{apps/v1 Deployment test-deployment 5fec78d8-938b-4640-bd31-571abea9033e 0xc00216a247 0xc00216a248}] [] [{kube-controller-manager Update apps/v1 2022-12-14 15:54:10 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"5fec78d8-938b-4640-bd31-571abea9033e\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 15:54:11 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*2,Selector:&v1.LabelSelector{MatchLabels:map[string]string{pod-template-hash: 7c7d8d58c8,test-deployment-static: true,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[] [] [] []} {[] [] [{test-deployment registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc00216a2e0 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:2,FullyLabeledReplicas:2,ObservedGeneration:2,ReadyReplicas:2,AvailableReplicas:2,Conditions:[]ReplicaSetCondition{},},} - - Dec 14 15:54:11.407: INFO: pod: "test-deployment-7c7d8d58c8-5fzd2": - &Pod{ObjectMeta:{test-deployment-7c7d8d58c8-5fzd2 test-deployment-7c7d8d58c8- deployment-1216 6fde8472-2be0-4711-8138-125cc950cbdb 6111 0 2022-12-14 15:54:09 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[] [{apps/v1 ReplicaSet test-deployment-7c7d8d58c8 05308084-3ea1-4062-9c29-bd6071e17afe 0xc002920a67 0xc002920a68}] [] [{kube-controller-manager Update v1 2022-12-14 15:54:09 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"05308084-3ea1-4062-9c29-bd6071e17afe\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 15:54:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.45\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-8tn6q,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:test-deployment,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8tn6q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*1,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:09 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.45,StartTime:2022-12-14 15:54:09 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:test-deployment,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 15:54:09 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://22f7a860436c766ba8a7c3133d35ada2acee6d7a15455a317366ca969516ce2e,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.45,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - - Dec 14 15:54:11.407: INFO: pod: "test-deployment-7c7d8d58c8-6kskv": - &Pod{ObjectMeta:{test-deployment-7c7d8d58c8-6kskv test-deployment-7c7d8d58c8- deployment-1216 d16be486-e0da-4f3c-86ac-291bfad58659 6146 0 2022-12-14 15:54:10 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[] [{apps/v1 ReplicaSet test-deployment-7c7d8d58c8 05308084-3ea1-4062-9c29-bd6071e17afe 0xc002920c67 0xc002920c68}] [] [{kube-controller-manager Update v1 2022-12-14 15:54:10 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"05308084-3ea1-4062-9c29-bd6071e17afe\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 15:54:11 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.13\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-2cltv,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:test-deployment,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2cltv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*1,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:11 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:11 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 15:54:10 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:10.233.64.13,StartTime:2022-12-14 15:54:10 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:test-deployment,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 15:54:10 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://8af110942d3d090b0ece07a95f0e9ab6e3e8d3d0c93508ace700d146d9d886b5,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.13,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - - Dec 14 15:54:11.408: INFO: ReplicaSet "test-deployment-8594bb6fdd": - &ReplicaSet{ObjectMeta:{test-deployment-8594bb6fdd deployment-1216 afb408ea-ad0a-4318-8204-af21e8672f0f 6074 3 2022-12-14 15:54:06 +0000 UTC map[pod-template-hash:8594bb6fdd test-deployment-static:true] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-deployment 5fec78d8-938b-4640-bd31-571abea9033e 0xc00216a347 0xc00216a348}] [] [{kube-controller-manager Update apps/v1 2022-12-14 15:54:08 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"5fec78d8-938b-4640-bd31-571abea9033e\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 15:54:08 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{pod-template-hash: 8594bb6fdd,test-deployment-static: true,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[pod-template-hash:8594bb6fdd test-deployment-static:true] map[] [] [] []} {[] [] [{test-deployment registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc00216a3d0 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} - - [AfterEach] [sig-apps] Deployment + STEP: Creating a kubernetes client 02/25/23 09:53:27.281 + Feb 25 09:53:27.281: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename var-expansion 02/25/23 09:53:27.284 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:53:27.311 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:53:27.315 + [It] should verify that a failing subpath expansion can be modified during the lifecycle of a container [Slow] [Conformance] + test/e2e/common/node/expansion.go:224 + STEP: creating the pod with failed condition 02/25/23 09:53:27.327 + Feb 25 09:53:27.346: INFO: Waiting up to 2m0s for pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92" in namespace "var-expansion-3821" to be "running" + Feb 25 09:53:27.353: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 7.395022ms + Feb 25 09:53:29.366: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019516304s + Feb 25 09:53:31.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 4.015993588s + Feb 25 09:53:33.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 6.016088713s + Feb 25 09:53:35.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 8.01645111s + Feb 25 09:53:37.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 10.014524243s + Feb 25 09:53:39.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 12.013531144s + Feb 25 09:53:41.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 14.015196458s + Feb 25 09:53:43.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 16.013769153s + Feb 25 09:53:45.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 18.014965925s + Feb 25 09:53:47.359: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 20.012613608s + Feb 25 09:53:49.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 22.017262996s + Feb 25 09:53:51.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 24.017538904s + Feb 25 09:53:53.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 26.015213962s + Feb 25 09:53:55.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 28.018218849s + Feb 25 09:53:57.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 30.018031713s + Feb 25 09:53:59.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 32.014731889s + Feb 25 09:54:01.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 34.017081872s + Feb 25 09:54:03.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 36.015902906s + Feb 25 09:54:05.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 38.013930985s + Feb 25 09:54:07.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 40.014905028s + Feb 25 09:54:09.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 42.015621935s + Feb 25 09:54:11.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 44.016516021s + Feb 25 09:54:13.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 46.017632401s + Feb 25 09:54:15.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 48.015988989s + Feb 25 09:54:17.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 50.016509165s + Feb 25 09:54:19.365: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 52.019334567s + Feb 25 09:54:21.365: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 54.019274123s + Feb 25 09:54:23.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 56.015533614s + Feb 25 09:54:25.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 58.017517181s + Feb 25 09:54:27.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m0.016979961s + Feb 25 09:54:29.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m2.017387252s + Feb 25 09:54:31.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m4.014784211s + Feb 25 09:54:33.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m6.016313129s + Feb 25 09:54:35.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m8.015697455s + Feb 25 09:54:37.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m10.016457409s + Feb 25 09:54:39.359: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m12.012987704s + Feb 25 09:54:41.359: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m14.013363148s + Feb 25 09:54:43.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m16.014454441s + Feb 25 09:54:45.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m18.0181947s + Feb 25 09:54:47.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m20.01418567s + Feb 25 09:54:49.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m22.01502728s + Feb 25 09:54:51.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m24.018020774s + Feb 25 09:54:53.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m26.01391261s + Feb 25 09:54:55.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m28.015461151s + Feb 25 09:54:57.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m30.013652303s + Feb 25 09:54:59.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m32.016909507s + Feb 25 09:55:01.364: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m34.018446007s + Feb 25 09:55:03.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m36.015210318s + Feb 25 09:55:05.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m38.015723197s + Feb 25 09:55:07.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m40.015123919s + Feb 25 09:55:09.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m42.014854669s + Feb 25 09:55:11.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m44.015474257s + Feb 25 09:55:13.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m46.015208531s + Feb 25 09:55:15.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m48.015788742s + Feb 25 09:55:17.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m50.013513734s + Feb 25 09:55:19.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m52.013515429s + Feb 25 09:55:21.362: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m54.015952129s + Feb 25 09:55:23.360: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m56.014315363s + Feb 25 09:55:25.361: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 1m58.0149816s + Feb 25 09:55:27.363: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 2m0.017308997s + Feb 25 09:55:27.372: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 2m0.026252215s + STEP: updating the pod 02/25/23 09:55:27.373 + Feb 25 09:55:27.898: INFO: Successfully updated pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92" + STEP: waiting for pod running 02/25/23 09:55:27.898 + Feb 25 09:55:27.899: INFO: Waiting up to 2m0s for pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92" in namespace "var-expansion-3821" to be "running" + Feb 25 09:55:27.908: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Pending", Reason="", readiness=false. Elapsed: 8.936321ms + Feb 25 09:55:29.917: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92": Phase="Running", Reason="", readiness=true. Elapsed: 2.018493087s + Feb 25 09:55:29.917: INFO: Pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92" satisfied condition "running" + STEP: deleting the pod gracefully 02/25/23 09:55:29.917 + Feb 25 09:55:29.918: INFO: Deleting pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92" in namespace "var-expansion-3821" + Feb 25 09:55:29.935: INFO: Wait up to 5m0s for pod "var-expansion-a09e5317-0bde-43b3-9194-02e3b105aa92" to be fully deleted + [AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 - Dec 14 15:54:11.414: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "deployment-1216" for this suite. 12/14/22 15:54:11.423 + Feb 25 09:56:01.953: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "var-expansion-3821" for this suite. 02/25/23 09:56:01.963 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSS +SS ------------------------------ -[sig-storage] Downward API volume - should provide container's memory limit [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:206 -[BeforeEach] [sig-storage] Downward API volume +[sig-node] Variable Expansion + should fail substituting values in a volume subpath with absolute path [Slow] [Conformance] + test/e2e/common/node/expansion.go:185 +[BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:54:11.435 -Dec 14 15:54:11.436: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 15:54:11.44 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:54:11.47 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:54:11.473 -[BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 -[It] should provide container's memory limit [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:206 -STEP: Creating a pod to test downward API volume plugin 12/14/22 15:54:11.477 -Dec 14 15:54:11.498: INFO: Waiting up to 5m0s for pod "downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7" in namespace "downward-api-4496" to be "Succeeded or Failed" -Dec 14 15:54:11.503: INFO: Pod "downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7": Phase="Pending", Reason="", readiness=false. Elapsed: 4.968579ms -Dec 14 15:54:13.513: INFO: Pod "downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014661969s -Dec 14 15:54:15.514: INFO: Pod "downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016113894s -STEP: Saw pod success 12/14/22 15:54:15.514 -Dec 14 15:54:15.515: INFO: Pod "downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7" satisfied condition "Succeeded or Failed" -Dec 14 15:54:15.522: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7 container client-container: -STEP: delete the pod 12/14/22 15:54:15.553 -Dec 14 15:54:15.577: INFO: Waiting for pod downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7 to disappear -Dec 14 15:54:15.584: INFO: Pod downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7 no longer exists -[AfterEach] [sig-storage] Downward API volume +STEP: Creating a kubernetes client 02/25/23 09:56:01.975 +Feb 25 09:56:01.975: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename var-expansion 02/25/23 09:56:01.978 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:56:02.005 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:56:02.009 +[It] should fail substituting values in a volume subpath with absolute path [Slow] [Conformance] + test/e2e/common/node/expansion.go:185 +Feb 25 09:56:02.026: INFO: Waiting up to 2m0s for pod "var-expansion-34f13bb3-f01d-4fe3-9bc0-e8b4fb28731f" in namespace "var-expansion-113" to be "container 0 failed with reason CreateContainerConfigError" +Feb 25 09:56:02.037: INFO: Pod "var-expansion-34f13bb3-f01d-4fe3-9bc0-e8b4fb28731f": Phase="Pending", Reason="", readiness=false. Elapsed: 10.96569ms +Feb 25 09:56:04.046: INFO: Pod "var-expansion-34f13bb3-f01d-4fe3-9bc0-e8b4fb28731f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019675564s +Feb 25 09:56:04.046: INFO: Pod "var-expansion-34f13bb3-f01d-4fe3-9bc0-e8b4fb28731f" satisfied condition "container 0 failed with reason CreateContainerConfigError" +Feb 25 09:56:04.046: INFO: Deleting pod "var-expansion-34f13bb3-f01d-4fe3-9bc0-e8b4fb28731f" in namespace "var-expansion-113" +Feb 25 09:56:04.060: INFO: Wait up to 5m0s for pod "var-expansion-34f13bb3-f01d-4fe3-9bc0-e8b4fb28731f" to be fully deleted +[AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 -Dec 14 15:54:15.585: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-4496" for this suite. 12/14/22 15:54:15.593 -{"msg":"PASSED [sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance]","completed":34,"skipped":512,"failed":0} +Feb 25 09:56:06.078: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "var-expansion-113" for this suite. 02/25/23 09:56:06.086 +{"msg":"PASSED [sig-node] Variable Expansion should fail substituting values in a volume subpath with absolute path [Slow] [Conformance]","completed":35,"skipped":897,"failed":0} ------------------------------ -• [4.174 seconds] -[sig-storage] Downward API volume -test/e2e/common/storage/framework.go:23 - should provide container's memory limit [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:206 +• [4.132 seconds] +[sig-node] Variable Expansion +test/e2e/common/node/framework.go:23 + should fail substituting values in a volume subpath with absolute path [Slow] [Conformance] + test/e2e/common/node/expansion.go:185 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Downward API volume + [BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:54:11.435 - Dec 14 15:54:11.436: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 15:54:11.44 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:54:11.47 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:54:11.473 - [BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 - [It] should provide container's memory limit [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:206 - STEP: Creating a pod to test downward API volume plugin 12/14/22 15:54:11.477 - Dec 14 15:54:11.498: INFO: Waiting up to 5m0s for pod "downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7" in namespace "downward-api-4496" to be "Succeeded or Failed" - Dec 14 15:54:11.503: INFO: Pod "downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7": Phase="Pending", Reason="", readiness=false. Elapsed: 4.968579ms - Dec 14 15:54:13.513: INFO: Pod "downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014661969s - Dec 14 15:54:15.514: INFO: Pod "downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016113894s - STEP: Saw pod success 12/14/22 15:54:15.514 - Dec 14 15:54:15.515: INFO: Pod "downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7" satisfied condition "Succeeded or Failed" - Dec 14 15:54:15.522: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7 container client-container: - STEP: delete the pod 12/14/22 15:54:15.553 - Dec 14 15:54:15.577: INFO: Waiting for pod downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7 to disappear - Dec 14 15:54:15.584: INFO: Pod downwardapi-volume-9b8ef978-8112-41eb-9bca-18a06a5b23f7 no longer exists - [AfterEach] [sig-storage] Downward API volume + STEP: Creating a kubernetes client 02/25/23 09:56:01.975 + Feb 25 09:56:01.975: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename var-expansion 02/25/23 09:56:01.978 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:56:02.005 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:56:02.009 + [It] should fail substituting values in a volume subpath with absolute path [Slow] [Conformance] + test/e2e/common/node/expansion.go:185 + Feb 25 09:56:02.026: INFO: Waiting up to 2m0s for pod "var-expansion-34f13bb3-f01d-4fe3-9bc0-e8b4fb28731f" in namespace "var-expansion-113" to be "container 0 failed with reason CreateContainerConfigError" + Feb 25 09:56:02.037: INFO: Pod "var-expansion-34f13bb3-f01d-4fe3-9bc0-e8b4fb28731f": Phase="Pending", Reason="", readiness=false. Elapsed: 10.96569ms + Feb 25 09:56:04.046: INFO: Pod "var-expansion-34f13bb3-f01d-4fe3-9bc0-e8b4fb28731f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019675564s + Feb 25 09:56:04.046: INFO: Pod "var-expansion-34f13bb3-f01d-4fe3-9bc0-e8b4fb28731f" satisfied condition "container 0 failed with reason CreateContainerConfigError" + Feb 25 09:56:04.046: INFO: Deleting pod "var-expansion-34f13bb3-f01d-4fe3-9bc0-e8b4fb28731f" in namespace "var-expansion-113" + Feb 25 09:56:04.060: INFO: Wait up to 5m0s for pod "var-expansion-34f13bb3-f01d-4fe3-9bc0-e8b4fb28731f" to be fully deleted + [AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 - Dec 14 15:54:15.585: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-4496" for this suite. 12/14/22 15:54:15.593 + Feb 25 09:56:06.078: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "var-expansion-113" for this suite. 02/25/23 09:56:06.086 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSS +SSSSSSSSS ------------------------------ -[sig-storage] ConfigMap - optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:239 -[BeforeEach] [sig-storage] ConfigMap +[sig-node] NoExecuteTaintManager Single Pod [Serial] + removing taint cancels eviction [Disruptive] [Conformance] + test/e2e/node/taints.go:289 +[BeforeEach] [sig-node] NoExecuteTaintManager Single Pod [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:54:15.618 -Dec 14 15:54:15.619: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename configmap 12/14/22 15:54:15.621 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:54:15.654 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:54:15.661 -[It] optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:239 -STEP: Creating configMap with name cm-test-opt-del-4d8ca34b-cfce-4cb5-bb8c-4c0279ecc923 12/14/22 15:54:15.673 -STEP: Creating configMap with name cm-test-opt-upd-2be15fe0-32a6-4dd1-b9f3-a7035e527553 12/14/22 15:54:15.684 -STEP: Creating the pod 12/14/22 15:54:15.69 -Dec 14 15:54:15.703: INFO: Waiting up to 5m0s for pod "pod-configmaps-d080f854-8e8f-4469-9fb4-70e46bf8e997" in namespace "configmap-1824" to be "running and ready" -Dec 14 15:54:15.706: INFO: Pod "pod-configmaps-d080f854-8e8f-4469-9fb4-70e46bf8e997": Phase="Pending", Reason="", readiness=false. Elapsed: 3.505417ms -Dec 14 15:54:15.706: INFO: The phase of Pod pod-configmaps-d080f854-8e8f-4469-9fb4-70e46bf8e997 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 15:54:17.715: INFO: Pod "pod-configmaps-d080f854-8e8f-4469-9fb4-70e46bf8e997": Phase="Running", Reason="", readiness=true. Elapsed: 2.012720278s -Dec 14 15:54:17.715: INFO: The phase of Pod pod-configmaps-d080f854-8e8f-4469-9fb4-70e46bf8e997 is Running (Ready = true) -Dec 14 15:54:17.715: INFO: Pod "pod-configmaps-d080f854-8e8f-4469-9fb4-70e46bf8e997" satisfied condition "running and ready" -STEP: Deleting configmap cm-test-opt-del-4d8ca34b-cfce-4cb5-bb8c-4c0279ecc923 12/14/22 15:54:17.757 -STEP: Updating configmap cm-test-opt-upd-2be15fe0-32a6-4dd1-b9f3-a7035e527553 12/14/22 15:54:17.769 -STEP: Creating configMap with name cm-test-opt-create-37e5e93b-391b-4b80-a62a-db59d077bfb7 12/14/22 15:54:17.775 -STEP: waiting to observe update in volume 12/14/22 15:54:17.781 -[AfterEach] [sig-storage] ConfigMap +STEP: Creating a kubernetes client 02/25/23 09:56:06.107 +Feb 25 09:56:06.107: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename taint-single-pod 02/25/23 09:56:06.109 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:56:06.134 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:56:06.139 +[BeforeEach] [sig-node] NoExecuteTaintManager Single Pod [Serial] + test/e2e/node/taints.go:166 +Feb 25 09:56:06.144: INFO: Waiting up to 1m0s for all nodes to be ready +Feb 25 09:57:06.195: INFO: Waiting for terminating namespaces to be deleted... +[It] removing taint cancels eviction [Disruptive] [Conformance] + test/e2e/node/taints.go:289 +Feb 25 09:57:06.202: INFO: Starting informer... +STEP: Starting pod... 02/25/23 09:57:06.202 +Feb 25 09:57:06.432: INFO: Pod is running on diek3dua9gay-3. Tainting Node +STEP: Trying to apply a taint on the Node 02/25/23 09:57:06.432 +STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 02/25/23 09:57:06.456 +STEP: Waiting short time to make sure Pod is queued for deletion 02/25/23 09:57:06.468 +Feb 25 09:57:06.468: INFO: Pod wasn't evicted. Proceeding +Feb 25 09:57:06.468: INFO: Removing taint from Node +STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 02/25/23 09:57:06.496 +STEP: Waiting some time to make sure that toleration time passed. 02/25/23 09:57:06.522 +Feb 25 09:58:21.523: INFO: Pod wasn't evicted. Test successful +[AfterEach] [sig-node] NoExecuteTaintManager Single Pod [Serial] test/e2e/framework/framework.go:187 -Dec 14 15:54:19.859: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-1824" for this suite. 12/14/22 15:54:19.865 -{"msg":"PASSED [sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance]","completed":35,"skipped":522,"failed":0} +Feb 25 09:58:21.525: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "taint-single-pod-7279" for this suite. 02/25/23 09:58:21.537 +{"msg":"PASSED [sig-node] NoExecuteTaintManager Single Pod [Serial] removing taint cancels eviction [Disruptive] [Conformance]","completed":36,"skipped":906,"failed":0} ------------------------------ -• [4.256 seconds] -[sig-storage] ConfigMap -test/e2e/common/storage/framework.go:23 - optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:239 +• [SLOW TEST] [135.442 seconds] +[sig-node] NoExecuteTaintManager Single Pod [Serial] +test/e2e/node/framework.go:23 + removing taint cancels eviction [Disruptive] [Conformance] + test/e2e/node/taints.go:289 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] ConfigMap + [BeforeEach] [sig-node] NoExecuteTaintManager Single Pod [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:54:15.618 - Dec 14 15:54:15.619: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename configmap 12/14/22 15:54:15.621 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:54:15.654 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:54:15.661 - [It] optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:239 - STEP: Creating configMap with name cm-test-opt-del-4d8ca34b-cfce-4cb5-bb8c-4c0279ecc923 12/14/22 15:54:15.673 - STEP: Creating configMap with name cm-test-opt-upd-2be15fe0-32a6-4dd1-b9f3-a7035e527553 12/14/22 15:54:15.684 - STEP: Creating the pod 12/14/22 15:54:15.69 - Dec 14 15:54:15.703: INFO: Waiting up to 5m0s for pod "pod-configmaps-d080f854-8e8f-4469-9fb4-70e46bf8e997" in namespace "configmap-1824" to be "running and ready" - Dec 14 15:54:15.706: INFO: Pod "pod-configmaps-d080f854-8e8f-4469-9fb4-70e46bf8e997": Phase="Pending", Reason="", readiness=false. Elapsed: 3.505417ms - Dec 14 15:54:15.706: INFO: The phase of Pod pod-configmaps-d080f854-8e8f-4469-9fb4-70e46bf8e997 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 15:54:17.715: INFO: Pod "pod-configmaps-d080f854-8e8f-4469-9fb4-70e46bf8e997": Phase="Running", Reason="", readiness=true. Elapsed: 2.012720278s - Dec 14 15:54:17.715: INFO: The phase of Pod pod-configmaps-d080f854-8e8f-4469-9fb4-70e46bf8e997 is Running (Ready = true) - Dec 14 15:54:17.715: INFO: Pod "pod-configmaps-d080f854-8e8f-4469-9fb4-70e46bf8e997" satisfied condition "running and ready" - STEP: Deleting configmap cm-test-opt-del-4d8ca34b-cfce-4cb5-bb8c-4c0279ecc923 12/14/22 15:54:17.757 - STEP: Updating configmap cm-test-opt-upd-2be15fe0-32a6-4dd1-b9f3-a7035e527553 12/14/22 15:54:17.769 - STEP: Creating configMap with name cm-test-opt-create-37e5e93b-391b-4b80-a62a-db59d077bfb7 12/14/22 15:54:17.775 - STEP: waiting to observe update in volume 12/14/22 15:54:17.781 - [AfterEach] [sig-storage] ConfigMap + STEP: Creating a kubernetes client 02/25/23 09:56:06.107 + Feb 25 09:56:06.107: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename taint-single-pod 02/25/23 09:56:06.109 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:56:06.134 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:56:06.139 + [BeforeEach] [sig-node] NoExecuteTaintManager Single Pod [Serial] + test/e2e/node/taints.go:166 + Feb 25 09:56:06.144: INFO: Waiting up to 1m0s for all nodes to be ready + Feb 25 09:57:06.195: INFO: Waiting for terminating namespaces to be deleted... + [It] removing taint cancels eviction [Disruptive] [Conformance] + test/e2e/node/taints.go:289 + Feb 25 09:57:06.202: INFO: Starting informer... + STEP: Starting pod... 02/25/23 09:57:06.202 + Feb 25 09:57:06.432: INFO: Pod is running on diek3dua9gay-3. Tainting Node + STEP: Trying to apply a taint on the Node 02/25/23 09:57:06.432 + STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 02/25/23 09:57:06.456 + STEP: Waiting short time to make sure Pod is queued for deletion 02/25/23 09:57:06.468 + Feb 25 09:57:06.468: INFO: Pod wasn't evicted. Proceeding + Feb 25 09:57:06.468: INFO: Removing taint from Node + STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 02/25/23 09:57:06.496 + STEP: Waiting some time to make sure that toleration time passed. 02/25/23 09:57:06.522 + Feb 25 09:58:21.523: INFO: Pod wasn't evicted. Test successful + [AfterEach] [sig-node] NoExecuteTaintManager Single Pod [Serial] test/e2e/framework/framework.go:187 - Dec 14 15:54:19.859: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "configmap-1824" for this suite. 12/14/22 15:54:19.865 + Feb 25 09:58:21.525: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "taint-single-pod-7279" for this suite. 02/25/23 09:58:21.537 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Probing container - with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:68 -[BeforeEach] [sig-node] Probing container +[sig-node] Security Context When creating a container with runAsUser + should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:346 +[BeforeEach] [sig-node] Security Context test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:54:19.877 -Dec 14 15:54:19.878: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-probe 12/14/22 15:54:19.88 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:54:19.907 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:54:19.91 -[BeforeEach] [sig-node] Probing container - test/e2e/common/node/container_probe.go:59 -[It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:68 -Dec 14 15:54:19.923: INFO: Waiting up to 5m0s for pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8" in namespace "container-probe-9644" to be "running and ready" -Dec 14 15:54:19.928: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Pending", Reason="", readiness=false. Elapsed: 5.00005ms -Dec 14 15:54:19.928: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 15:54:21.938: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 2.015075716s -Dec 14 15:54:21.938: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) -Dec 14 15:54:23.936: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 4.013256859s -Dec 14 15:54:23.936: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) -Dec 14 15:54:25.937: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 6.014236189s -Dec 14 15:54:25.937: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) -Dec 14 15:54:27.936: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 8.012843021s -Dec 14 15:54:27.936: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) -Dec 14 15:54:29.939: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 10.016080054s -Dec 14 15:54:29.939: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) -Dec 14 15:54:31.935: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 12.01253667s -Dec 14 15:54:31.935: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) -Dec 14 15:54:33.938: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 14.015746736s -Dec 14 15:54:33.939: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) -Dec 14 15:54:35.940: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 16.017587761s -Dec 14 15:54:35.941: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) -Dec 14 15:54:37.939: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 18.016446479s -Dec 14 15:54:37.940: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) -Dec 14 15:54:39.948: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 20.025715732s -Dec 14 15:54:39.949: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) -Dec 14 15:54:41.938: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=true. Elapsed: 22.01535441s -Dec 14 15:54:41.938: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = true) -Dec 14 15:54:41.938: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8" satisfied condition "running and ready" -Dec 14 15:54:41.944: INFO: Container started at 2022-12-14 15:54:20 +0000 UTC, pod became ready at 2022-12-14 15:54:40 +0000 UTC -[AfterEach] [sig-node] Probing container +STEP: Creating a kubernetes client 02/25/23 09:58:21.561 +Feb 25 09:58:21.561: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename security-context-test 02/25/23 09:58:21.565 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:58:21.632 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:58:21.636 +[BeforeEach] [sig-node] Security Context + test/e2e/common/node/security_context.go:49 +[It] should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:346 +Feb 25 09:58:21.654: INFO: Waiting up to 5m0s for pod "busybox-user-65534-60c71909-94fa-4ca0-aa8d-4fd4bb3ee48d" in namespace "security-context-test-3282" to be "Succeeded or Failed" +Feb 25 09:58:21.698: INFO: Pod "busybox-user-65534-60c71909-94fa-4ca0-aa8d-4fd4bb3ee48d": Phase="Pending", Reason="", readiness=false. Elapsed: 44.068361ms +Feb 25 09:58:23.706: INFO: Pod "busybox-user-65534-60c71909-94fa-4ca0-aa8d-4fd4bb3ee48d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.051765718s +Feb 25 09:58:25.707: INFO: Pod "busybox-user-65534-60c71909-94fa-4ca0-aa8d-4fd4bb3ee48d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.05304435s +Feb 25 09:58:25.707: INFO: Pod "busybox-user-65534-60c71909-94fa-4ca0-aa8d-4fd4bb3ee48d" satisfied condition "Succeeded or Failed" +[AfterEach] [sig-node] Security Context test/e2e/framework/framework.go:187 -Dec 14 15:54:41.945: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-probe-9644" for this suite. 12/14/22 15:54:41.954 -{"msg":"PASSED [sig-node] Probing container with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance]","completed":36,"skipped":553,"failed":0} +Feb 25 09:58:25.708: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "security-context-test-3282" for this suite. 02/25/23 09:58:25.715 +{"msg":"PASSED [sig-node] Security Context When creating a container with runAsUser should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]","completed":37,"skipped":928,"failed":0} ------------------------------ -• [SLOW TEST] [22.088 seconds] -[sig-node] Probing container +• [4.168 seconds] +[sig-node] Security Context test/e2e/common/node/framework.go:23 - with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:68 + When creating a container with runAsUser + test/e2e/common/node/security_context.go:308 + should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:346 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Probing container + [BeforeEach] [sig-node] Security Context test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:54:19.877 - Dec 14 15:54:19.878: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-probe 12/14/22 15:54:19.88 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:54:19.907 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:54:19.91 - [BeforeEach] [sig-node] Probing container - test/e2e/common/node/container_probe.go:59 - [It] with readiness probe should not be ready before initial delay and never restart [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:68 - Dec 14 15:54:19.923: INFO: Waiting up to 5m0s for pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8" in namespace "container-probe-9644" to be "running and ready" - Dec 14 15:54:19.928: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Pending", Reason="", readiness=false. Elapsed: 5.00005ms - Dec 14 15:54:19.928: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 15:54:21.938: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 2.015075716s - Dec 14 15:54:21.938: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) - Dec 14 15:54:23.936: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 4.013256859s - Dec 14 15:54:23.936: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) - Dec 14 15:54:25.937: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 6.014236189s - Dec 14 15:54:25.937: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) - Dec 14 15:54:27.936: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 8.012843021s - Dec 14 15:54:27.936: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) - Dec 14 15:54:29.939: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 10.016080054s - Dec 14 15:54:29.939: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) - Dec 14 15:54:31.935: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 12.01253667s - Dec 14 15:54:31.935: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) - Dec 14 15:54:33.938: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 14.015746736s - Dec 14 15:54:33.939: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) - Dec 14 15:54:35.940: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 16.017587761s - Dec 14 15:54:35.941: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) - Dec 14 15:54:37.939: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 18.016446479s - Dec 14 15:54:37.940: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) - Dec 14 15:54:39.948: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=false. Elapsed: 20.025715732s - Dec 14 15:54:39.949: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = false) - Dec 14 15:54:41.938: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8": Phase="Running", Reason="", readiness=true. Elapsed: 22.01535441s - Dec 14 15:54:41.938: INFO: The phase of Pod test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8 is Running (Ready = true) - Dec 14 15:54:41.938: INFO: Pod "test-webserver-6f5adba4-aa10-4049-948b-2ee0c78d87f8" satisfied condition "running and ready" - Dec 14 15:54:41.944: INFO: Container started at 2022-12-14 15:54:20 +0000 UTC, pod became ready at 2022-12-14 15:54:40 +0000 UTC - [AfterEach] [sig-node] Probing container + STEP: Creating a kubernetes client 02/25/23 09:58:21.561 + Feb 25 09:58:21.561: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename security-context-test 02/25/23 09:58:21.565 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:58:21.632 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:58:21.636 + [BeforeEach] [sig-node] Security Context + test/e2e/common/node/security_context.go:49 + [It] should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:346 + Feb 25 09:58:21.654: INFO: Waiting up to 5m0s for pod "busybox-user-65534-60c71909-94fa-4ca0-aa8d-4fd4bb3ee48d" in namespace "security-context-test-3282" to be "Succeeded or Failed" + Feb 25 09:58:21.698: INFO: Pod "busybox-user-65534-60c71909-94fa-4ca0-aa8d-4fd4bb3ee48d": Phase="Pending", Reason="", readiness=false. Elapsed: 44.068361ms + Feb 25 09:58:23.706: INFO: Pod "busybox-user-65534-60c71909-94fa-4ca0-aa8d-4fd4bb3ee48d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.051765718s + Feb 25 09:58:25.707: INFO: Pod "busybox-user-65534-60c71909-94fa-4ca0-aa8d-4fd4bb3ee48d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.05304435s + Feb 25 09:58:25.707: INFO: Pod "busybox-user-65534-60c71909-94fa-4ca0-aa8d-4fd4bb3ee48d" satisfied condition "Succeeded or Failed" + [AfterEach] [sig-node] Security Context test/e2e/framework/framework.go:187 - Dec 14 15:54:41.945: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-probe-9644" for this suite. 12/14/22 15:54:41.954 + Feb 25 09:58:25.708: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "security-context-test-3282" for this suite. 02/25/23 09:58:25.715 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] Garbage collector - should orphan pods created by rc if delete options say so [Conformance] - test/e2e/apimachinery/garbage_collector.go:370 -[BeforeEach] [sig-api-machinery] Garbage collector +[sig-storage] Projected downwardAPI + should provide container's memory request [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:234 +[BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:54:41.978 -Dec 14 15:54:41.979: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename gc 12/14/22 15:54:41.983 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:54:42.009 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:54:42.013 -[It] should orphan pods created by rc if delete options say so [Conformance] - test/e2e/apimachinery/garbage_collector.go:370 -STEP: create the rc 12/14/22 15:54:42.023 -STEP: delete the rc 12/14/22 15:54:47.041 -STEP: wait for the rc to be deleted 12/14/22 15:54:47.227 -STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods 12/14/22 15:54:52.234 -STEP: Gathering metrics 12/14/22 15:55:22.272 -Dec 14 15:55:22.315: INFO: Waiting up to 5m0s for pod "kube-controller-manager-kiedai4iakae-2" in namespace "kube-system" to be "running and ready" -Dec 14 15:55:22.321: INFO: Pod "kube-controller-manager-kiedai4iakae-2": Phase="Running", Reason="", readiness=true. Elapsed: 6.67983ms -Dec 14 15:55:22.322: INFO: The phase of Pod kube-controller-manager-kiedai4iakae-2 is Running (Ready = true) -Dec 14 15:55:22.322: INFO: Pod "kube-controller-manager-kiedai4iakae-2" satisfied condition "running and ready" -Dec 14 15:55:22.438: INFO: For apiserver_request_total: -For apiserver_request_latency_seconds: -For apiserver_init_events_total: -For garbage_collector_attempt_to_delete_queue_latency: -For garbage_collector_attempt_to_delete_work_duration: -For garbage_collector_attempt_to_orphan_queue_latency: -For garbage_collector_attempt_to_orphan_work_duration: -For garbage_collector_dirty_processing_latency_microseconds: -For garbage_collector_event_processing_latency_microseconds: -For garbage_collector_graph_changes_queue_latency: -For garbage_collector_graph_changes_work_duration: -For garbage_collector_orphan_processing_latency_microseconds: -For namespace_queue_latency: -For namespace_queue_latency_sum: -For namespace_queue_latency_count: -For namespace_retries: -For namespace_work_duration: -For namespace_work_duration_sum: -For namespace_work_duration_count: -For function_duration_seconds: -For errors_total: -For evicted_pods_total: - -Dec 14 15:55:22.438: INFO: Deleting pod "simpletest.rc-28jn8" in namespace "gc-2522" -Dec 14 15:55:22.460: INFO: Deleting pod "simpletest.rc-2vsvg" in namespace "gc-2522" -Dec 14 15:55:22.516: INFO: Deleting pod "simpletest.rc-44dqg" in namespace "gc-2522" -Dec 14 15:55:22.564: INFO: Deleting pod "simpletest.rc-47zdm" in namespace "gc-2522" -Dec 14 15:55:22.622: INFO: Deleting pod "simpletest.rc-4sstw" in namespace "gc-2522" -Dec 14 15:55:22.642: INFO: Deleting pod "simpletest.rc-54vvm" in namespace "gc-2522" -Dec 14 15:55:22.673: INFO: Deleting pod "simpletest.rc-5tnst" in namespace "gc-2522" -Dec 14 15:55:22.694: INFO: Deleting pod "simpletest.rc-5tqjd" in namespace "gc-2522" -Dec 14 15:55:22.764: INFO: Deleting pod "simpletest.rc-695px" in namespace "gc-2522" -Dec 14 15:55:22.819: INFO: Deleting pod "simpletest.rc-6d2tj" in namespace "gc-2522" -Dec 14 15:55:22.857: INFO: Deleting pod "simpletest.rc-6jktk" in namespace "gc-2522" -Dec 14 15:55:22.920: INFO: Deleting pod "simpletest.rc-6zf77" in namespace "gc-2522" -Dec 14 15:55:22.958: INFO: Deleting pod "simpletest.rc-7q98c" in namespace "gc-2522" -Dec 14 15:55:23.010: INFO: Deleting pod "simpletest.rc-7vsft" in namespace "gc-2522" -Dec 14 15:55:23.068: INFO: Deleting pod "simpletest.rc-87mqj" in namespace "gc-2522" -Dec 14 15:55:23.233: INFO: Deleting pod "simpletest.rc-8pg6g" in namespace "gc-2522" -Dec 14 15:55:23.324: INFO: Deleting pod "simpletest.rc-928k2" in namespace "gc-2522" -Dec 14 15:55:23.366: INFO: Deleting pod "simpletest.rc-972cc" in namespace "gc-2522" -Dec 14 15:55:23.462: INFO: Deleting pod "simpletest.rc-98rbc" in namespace "gc-2522" -Dec 14 15:55:23.497: INFO: Deleting pod "simpletest.rc-9cf6x" in namespace "gc-2522" -Dec 14 15:55:23.549: INFO: Deleting pod "simpletest.rc-9m9cb" in namespace "gc-2522" -Dec 14 15:55:23.580: INFO: Deleting pod "simpletest.rc-9xgq9" in namespace "gc-2522" -Dec 14 15:55:23.621: INFO: Deleting pod "simpletest.rc-b5qxw" in namespace "gc-2522" -Dec 14 15:55:23.671: INFO: Deleting pod "simpletest.rc-bbz4j" in namespace "gc-2522" -Dec 14 15:55:23.704: INFO: Deleting pod "simpletest.rc-bc9t5" in namespace "gc-2522" -Dec 14 15:55:23.727: INFO: Deleting pod "simpletest.rc-bfszv" in namespace "gc-2522" -Dec 14 15:55:23.765: INFO: Deleting pod "simpletest.rc-bl98x" in namespace "gc-2522" -Dec 14 15:55:23.831: INFO: Deleting pod "simpletest.rc-c8t5g" in namespace "gc-2522" -Dec 14 15:55:23.879: INFO: Deleting pod "simpletest.rc-csn8z" in namespace "gc-2522" -Dec 14 15:55:23.931: INFO: Deleting pod "simpletest.rc-cswg4" in namespace "gc-2522" -Dec 14 15:55:23.983: INFO: Deleting pod "simpletest.rc-d4nvb" in namespace "gc-2522" -Dec 14 15:55:24.013: INFO: Deleting pod "simpletest.rc-d6fdk" in namespace "gc-2522" -Dec 14 15:55:24.048: INFO: Deleting pod "simpletest.rc-d96xw" in namespace "gc-2522" -Dec 14 15:55:24.071: INFO: Deleting pod "simpletest.rc-d9k9w" in namespace "gc-2522" -Dec 14 15:55:24.101: INFO: Deleting pod "simpletest.rc-dn27j" in namespace "gc-2522" -Dec 14 15:55:24.182: INFO: Deleting pod "simpletest.rc-dvskv" in namespace "gc-2522" -Dec 14 15:55:24.217: INFO: Deleting pod "simpletest.rc-fdxmq" in namespace "gc-2522" -Dec 14 15:55:24.305: INFO: Deleting pod "simpletest.rc-ffz4k" in namespace "gc-2522" -Dec 14 15:55:24.340: INFO: Deleting pod "simpletest.rc-fg4cc" in namespace "gc-2522" -Dec 14 15:55:24.368: INFO: Deleting pod "simpletest.rc-fj8bx" in namespace "gc-2522" -Dec 14 15:55:24.395: INFO: Deleting pod "simpletest.rc-fjnc7" in namespace "gc-2522" -Dec 14 15:55:24.445: INFO: Deleting pod "simpletest.rc-fnwfj" in namespace "gc-2522" -Dec 14 15:55:24.491: INFO: Deleting pod "simpletest.rc-fx8w4" in namespace "gc-2522" -Dec 14 15:55:24.549: INFO: Deleting pod "simpletest.rc-g4g92" in namespace "gc-2522" -Dec 14 15:55:24.635: INFO: Deleting pod "simpletest.rc-g8j4f" in namespace "gc-2522" -Dec 14 15:55:24.687: INFO: Deleting pod "simpletest.rc-g9ft4" in namespace "gc-2522" -Dec 14 15:55:24.733: INFO: Deleting pod "simpletest.rc-hx4nq" in namespace "gc-2522" -Dec 14 15:55:24.797: INFO: Deleting pod "simpletest.rc-j6g4w" in namespace "gc-2522" -Dec 14 15:55:24.879: INFO: Deleting pod "simpletest.rc-jbmhm" in namespace "gc-2522" -Dec 14 15:55:24.935: INFO: Deleting pod "simpletest.rc-jgbrr" in namespace "gc-2522" -Dec 14 15:55:24.989: INFO: Deleting pod "simpletest.rc-jqvbz" in namespace "gc-2522" -Dec 14 15:55:25.039: INFO: Deleting pod "simpletest.rc-jthff" in namespace "gc-2522" -Dec 14 15:55:25.148: INFO: Deleting pod "simpletest.rc-k227x" in namespace "gc-2522" -Dec 14 15:55:25.288: INFO: Deleting pod "simpletest.rc-k7fc8" in namespace "gc-2522" -Dec 14 15:55:25.390: INFO: Deleting pod "simpletest.rc-k8rzb" in namespace "gc-2522" -Dec 14 15:55:25.434: INFO: Deleting pod "simpletest.rc-kng8x" in namespace "gc-2522" -Dec 14 15:55:25.471: INFO: Deleting pod "simpletest.rc-kw9jq" in namespace "gc-2522" -Dec 14 15:55:25.517: INFO: Deleting pod "simpletest.rc-l5p4v" in namespace "gc-2522" -Dec 14 15:55:25.566: INFO: Deleting pod "simpletest.rc-lhdb4" in namespace "gc-2522" -Dec 14 15:55:25.618: INFO: Deleting pod "simpletest.rc-lhldf" in namespace "gc-2522" -Dec 14 15:55:25.654: INFO: Deleting pod "simpletest.rc-lj5rk" in namespace "gc-2522" -Dec 14 15:55:25.691: INFO: Deleting pod "simpletest.rc-mp5x2" in namespace "gc-2522" -Dec 14 15:55:25.718: INFO: Deleting pod "simpletest.rc-n4ct4" in namespace "gc-2522" -Dec 14 15:55:25.742: INFO: Deleting pod "simpletest.rc-n7lj7" in namespace "gc-2522" -Dec 14 15:55:25.760: INFO: Deleting pod "simpletest.rc-nb4m6" in namespace "gc-2522" -Dec 14 15:55:25.824: INFO: Deleting pod "simpletest.rc-nx4nq" in namespace "gc-2522" -Dec 14 15:55:25.854: INFO: Deleting pod "simpletest.rc-p8jbf" in namespace "gc-2522" -Dec 14 15:55:25.886: INFO: Deleting pod "simpletest.rc-pcqvj" in namespace "gc-2522" -Dec 14 15:55:25.934: INFO: Deleting pod "simpletest.rc-pdgxn" in namespace "gc-2522" -Dec 14 15:55:25.959: INFO: Deleting pod "simpletest.rc-pwqx6" in namespace "gc-2522" -Dec 14 15:55:25.980: INFO: Deleting pod "simpletest.rc-q67qc" in namespace "gc-2522" -Dec 14 15:55:26.026: INFO: Deleting pod "simpletest.rc-qfhg7" in namespace "gc-2522" -Dec 14 15:55:26.068: INFO: Deleting pod "simpletest.rc-qj7lr" in namespace "gc-2522" -Dec 14 15:55:26.097: INFO: Deleting pod "simpletest.rc-qrfkp" in namespace "gc-2522" -Dec 14 15:55:26.131: INFO: Deleting pod "simpletest.rc-rc7vf" in namespace "gc-2522" -Dec 14 15:55:26.176: INFO: Deleting pod "simpletest.rc-rd6xv" in namespace "gc-2522" -Dec 14 15:55:26.210: INFO: Deleting pod "simpletest.rc-rdhdv" in namespace "gc-2522" -Dec 14 15:55:26.268: INFO: Deleting pod "simpletest.rc-sl2mp" in namespace "gc-2522" -Dec 14 15:55:26.303: INFO: Deleting pod "simpletest.rc-tbfzg" in namespace "gc-2522" -Dec 14 15:55:26.345: INFO: Deleting pod "simpletest.rc-td7j2" in namespace "gc-2522" -Dec 14 15:55:26.396: INFO: Deleting pod "simpletest.rc-tlk5d" in namespace "gc-2522" -Dec 14 15:55:26.444: INFO: Deleting pod "simpletest.rc-tmlb9" in namespace "gc-2522" -Dec 14 15:55:26.506: INFO: Deleting pod "simpletest.rc-v2dqg" in namespace "gc-2522" -Dec 14 15:55:26.522: INFO: Deleting pod "simpletest.rc-v9gjf" in namespace "gc-2522" -Dec 14 15:55:26.550: INFO: Deleting pod "simpletest.rc-vgvqx" in namespace "gc-2522" -Dec 14 15:55:26.619: INFO: Deleting pod "simpletest.rc-vzqct" in namespace "gc-2522" -Dec 14 15:55:26.696: INFO: Deleting pod "simpletest.rc-w5hnj" in namespace "gc-2522" -Dec 14 15:55:26.731: INFO: Deleting pod "simpletest.rc-w5tl9" in namespace "gc-2522" -Dec 14 15:55:26.751: INFO: Deleting pod "simpletest.rc-w6m5b" in namespace "gc-2522" -Dec 14 15:55:26.779: INFO: Deleting pod "simpletest.rc-wkfqf" in namespace "gc-2522" -Dec 14 15:55:26.821: INFO: Deleting pod "simpletest.rc-xgh98" in namespace "gc-2522" -Dec 14 15:55:26.854: INFO: Deleting pod "simpletest.rc-xj4dr" in namespace "gc-2522" -Dec 14 15:55:26.880: INFO: Deleting pod "simpletest.rc-xmkk4" in namespace "gc-2522" -Dec 14 15:55:26.923: INFO: Deleting pod "simpletest.rc-xs97t" in namespace "gc-2522" -Dec 14 15:55:26.948: INFO: Deleting pod "simpletest.rc-xsns8" in namespace "gc-2522" -Dec 14 15:55:26.980: INFO: Deleting pod "simpletest.rc-xwk68" in namespace "gc-2522" -Dec 14 15:55:27.020: INFO: Deleting pod "simpletest.rc-z49k8" in namespace "gc-2522" -Dec 14 15:55:27.048: INFO: Deleting pod "simpletest.rc-z5mgj" in namespace "gc-2522" -Dec 14 15:55:27.088: INFO: Deleting pod "simpletest.rc-z99b2" in namespace "gc-2522" -Dec 14 15:55:27.300: INFO: Deleting pod "simpletest.rc-zgmml" in namespace "gc-2522" -[AfterEach] [sig-api-machinery] Garbage collector +STEP: Creating a kubernetes client 02/25/23 09:58:25.74 +Feb 25 09:58:25.741: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 09:58:25.743 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:58:25.825 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:58:25.83 +[BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 +[It] should provide container's memory request [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:234 +STEP: Creating a pod to test downward API volume plugin 02/25/23 09:58:25.836 +Feb 25 09:58:25.859: INFO: Waiting up to 5m0s for pod "downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66" in namespace "projected-9269" to be "Succeeded or Failed" +Feb 25 09:58:25.879: INFO: Pod "downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66": Phase="Pending", Reason="", readiness=false. Elapsed: 19.192158ms +Feb 25 09:58:27.886: INFO: Pod "downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026824685s +Feb 25 09:58:29.888: INFO: Pod "downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.028054387s +STEP: Saw pod success 02/25/23 09:58:29.888 +Feb 25 09:58:29.889: INFO: Pod "downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66" satisfied condition "Succeeded or Failed" +Feb 25 09:58:29.895: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66 container client-container: +STEP: delete the pod 02/25/23 09:58:29.922 +Feb 25 09:58:29.945: INFO: Waiting for pod downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66 to disappear +Feb 25 09:58:29.952: INFO: Pod downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66 no longer exists +[AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 -Dec 14 15:55:27.375: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "gc-2522" for this suite. 12/14/22 15:55:27.382 -{"msg":"PASSED [sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Conformance]","completed":37,"skipped":585,"failed":0} +Feb 25 09:58:29.953: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-9269" for this suite. 02/25/23 09:58:29.961 +{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory request [NodeConformance] [Conformance]","completed":38,"skipped":980,"failed":0} ------------------------------ -• [SLOW TEST] [45.423 seconds] -[sig-api-machinery] Garbage collector -test/e2e/apimachinery/framework.go:23 - should orphan pods created by rc if delete options say so [Conformance] - test/e2e/apimachinery/garbage_collector.go:370 +• [4.235 seconds] +[sig-storage] Projected downwardAPI +test/e2e/common/storage/framework.go:23 + should provide container's memory request [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:234 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Garbage collector + [BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:54:41.978 - Dec 14 15:54:41.979: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename gc 12/14/22 15:54:41.983 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:54:42.009 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:54:42.013 - [It] should orphan pods created by rc if delete options say so [Conformance] - test/e2e/apimachinery/garbage_collector.go:370 - STEP: create the rc 12/14/22 15:54:42.023 - STEP: delete the rc 12/14/22 15:54:47.041 - STEP: wait for the rc to be deleted 12/14/22 15:54:47.227 - STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods 12/14/22 15:54:52.234 - STEP: Gathering metrics 12/14/22 15:55:22.272 - Dec 14 15:55:22.315: INFO: Waiting up to 5m0s for pod "kube-controller-manager-kiedai4iakae-2" in namespace "kube-system" to be "running and ready" - Dec 14 15:55:22.321: INFO: Pod "kube-controller-manager-kiedai4iakae-2": Phase="Running", Reason="", readiness=true. Elapsed: 6.67983ms - Dec 14 15:55:22.322: INFO: The phase of Pod kube-controller-manager-kiedai4iakae-2 is Running (Ready = true) - Dec 14 15:55:22.322: INFO: Pod "kube-controller-manager-kiedai4iakae-2" satisfied condition "running and ready" - Dec 14 15:55:22.438: INFO: For apiserver_request_total: - For apiserver_request_latency_seconds: - For apiserver_init_events_total: - For garbage_collector_attempt_to_delete_queue_latency: - For garbage_collector_attempt_to_delete_work_duration: - For garbage_collector_attempt_to_orphan_queue_latency: - For garbage_collector_attempt_to_orphan_work_duration: - For garbage_collector_dirty_processing_latency_microseconds: - For garbage_collector_event_processing_latency_microseconds: - For garbage_collector_graph_changes_queue_latency: - For garbage_collector_graph_changes_work_duration: - For garbage_collector_orphan_processing_latency_microseconds: - For namespace_queue_latency: - For namespace_queue_latency_sum: - For namespace_queue_latency_count: - For namespace_retries: - For namespace_work_duration: - For namespace_work_duration_sum: - For namespace_work_duration_count: - For function_duration_seconds: - For errors_total: - For evicted_pods_total: - - Dec 14 15:55:22.438: INFO: Deleting pod "simpletest.rc-28jn8" in namespace "gc-2522" - Dec 14 15:55:22.460: INFO: Deleting pod "simpletest.rc-2vsvg" in namespace "gc-2522" - Dec 14 15:55:22.516: INFO: Deleting pod "simpletest.rc-44dqg" in namespace "gc-2522" - Dec 14 15:55:22.564: INFO: Deleting pod "simpletest.rc-47zdm" in namespace "gc-2522" - Dec 14 15:55:22.622: INFO: Deleting pod "simpletest.rc-4sstw" in namespace "gc-2522" - Dec 14 15:55:22.642: INFO: Deleting pod "simpletest.rc-54vvm" in namespace "gc-2522" - Dec 14 15:55:22.673: INFO: Deleting pod "simpletest.rc-5tnst" in namespace "gc-2522" - Dec 14 15:55:22.694: INFO: Deleting pod "simpletest.rc-5tqjd" in namespace "gc-2522" - Dec 14 15:55:22.764: INFO: Deleting pod "simpletest.rc-695px" in namespace "gc-2522" - Dec 14 15:55:22.819: INFO: Deleting pod "simpletest.rc-6d2tj" in namespace "gc-2522" - Dec 14 15:55:22.857: INFO: Deleting pod "simpletest.rc-6jktk" in namespace "gc-2522" - Dec 14 15:55:22.920: INFO: Deleting pod "simpletest.rc-6zf77" in namespace "gc-2522" - Dec 14 15:55:22.958: INFO: Deleting pod "simpletest.rc-7q98c" in namespace "gc-2522" - Dec 14 15:55:23.010: INFO: Deleting pod "simpletest.rc-7vsft" in namespace "gc-2522" - Dec 14 15:55:23.068: INFO: Deleting pod "simpletest.rc-87mqj" in namespace "gc-2522" - Dec 14 15:55:23.233: INFO: Deleting pod "simpletest.rc-8pg6g" in namespace "gc-2522" - Dec 14 15:55:23.324: INFO: Deleting pod "simpletest.rc-928k2" in namespace "gc-2522" - Dec 14 15:55:23.366: INFO: Deleting pod "simpletest.rc-972cc" in namespace "gc-2522" - Dec 14 15:55:23.462: INFO: Deleting pod "simpletest.rc-98rbc" in namespace "gc-2522" - Dec 14 15:55:23.497: INFO: Deleting pod "simpletest.rc-9cf6x" in namespace "gc-2522" - Dec 14 15:55:23.549: INFO: Deleting pod "simpletest.rc-9m9cb" in namespace "gc-2522" - Dec 14 15:55:23.580: INFO: Deleting pod "simpletest.rc-9xgq9" in namespace "gc-2522" - Dec 14 15:55:23.621: INFO: Deleting pod "simpletest.rc-b5qxw" in namespace "gc-2522" - Dec 14 15:55:23.671: INFO: Deleting pod "simpletest.rc-bbz4j" in namespace "gc-2522" - Dec 14 15:55:23.704: INFO: Deleting pod "simpletest.rc-bc9t5" in namespace "gc-2522" - Dec 14 15:55:23.727: INFO: Deleting pod "simpletest.rc-bfszv" in namespace "gc-2522" - Dec 14 15:55:23.765: INFO: Deleting pod "simpletest.rc-bl98x" in namespace "gc-2522" - Dec 14 15:55:23.831: INFO: Deleting pod "simpletest.rc-c8t5g" in namespace "gc-2522" - Dec 14 15:55:23.879: INFO: Deleting pod "simpletest.rc-csn8z" in namespace "gc-2522" - Dec 14 15:55:23.931: INFO: Deleting pod "simpletest.rc-cswg4" in namespace "gc-2522" - Dec 14 15:55:23.983: INFO: Deleting pod "simpletest.rc-d4nvb" in namespace "gc-2522" - Dec 14 15:55:24.013: INFO: Deleting pod "simpletest.rc-d6fdk" in namespace "gc-2522" - Dec 14 15:55:24.048: INFO: Deleting pod "simpletest.rc-d96xw" in namespace "gc-2522" - Dec 14 15:55:24.071: INFO: Deleting pod "simpletest.rc-d9k9w" in namespace "gc-2522" - Dec 14 15:55:24.101: INFO: Deleting pod "simpletest.rc-dn27j" in namespace "gc-2522" - Dec 14 15:55:24.182: INFO: Deleting pod "simpletest.rc-dvskv" in namespace "gc-2522" - Dec 14 15:55:24.217: INFO: Deleting pod "simpletest.rc-fdxmq" in namespace "gc-2522" - Dec 14 15:55:24.305: INFO: Deleting pod "simpletest.rc-ffz4k" in namespace "gc-2522" - Dec 14 15:55:24.340: INFO: Deleting pod "simpletest.rc-fg4cc" in namespace "gc-2522" - Dec 14 15:55:24.368: INFO: Deleting pod "simpletest.rc-fj8bx" in namespace "gc-2522" - Dec 14 15:55:24.395: INFO: Deleting pod "simpletest.rc-fjnc7" in namespace "gc-2522" - Dec 14 15:55:24.445: INFO: Deleting pod "simpletest.rc-fnwfj" in namespace "gc-2522" - Dec 14 15:55:24.491: INFO: Deleting pod "simpletest.rc-fx8w4" in namespace "gc-2522" - Dec 14 15:55:24.549: INFO: Deleting pod "simpletest.rc-g4g92" in namespace "gc-2522" - Dec 14 15:55:24.635: INFO: Deleting pod "simpletest.rc-g8j4f" in namespace "gc-2522" - Dec 14 15:55:24.687: INFO: Deleting pod "simpletest.rc-g9ft4" in namespace "gc-2522" - Dec 14 15:55:24.733: INFO: Deleting pod "simpletest.rc-hx4nq" in namespace "gc-2522" - Dec 14 15:55:24.797: INFO: Deleting pod "simpletest.rc-j6g4w" in namespace "gc-2522" - Dec 14 15:55:24.879: INFO: Deleting pod "simpletest.rc-jbmhm" in namespace "gc-2522" - Dec 14 15:55:24.935: INFO: Deleting pod "simpletest.rc-jgbrr" in namespace "gc-2522" - Dec 14 15:55:24.989: INFO: Deleting pod "simpletest.rc-jqvbz" in namespace "gc-2522" - Dec 14 15:55:25.039: INFO: Deleting pod "simpletest.rc-jthff" in namespace "gc-2522" - Dec 14 15:55:25.148: INFO: Deleting pod "simpletest.rc-k227x" in namespace "gc-2522" - Dec 14 15:55:25.288: INFO: Deleting pod "simpletest.rc-k7fc8" in namespace "gc-2522" - Dec 14 15:55:25.390: INFO: Deleting pod "simpletest.rc-k8rzb" in namespace "gc-2522" - Dec 14 15:55:25.434: INFO: Deleting pod "simpletest.rc-kng8x" in namespace "gc-2522" - Dec 14 15:55:25.471: INFO: Deleting pod "simpletest.rc-kw9jq" in namespace "gc-2522" - Dec 14 15:55:25.517: INFO: Deleting pod "simpletest.rc-l5p4v" in namespace "gc-2522" - Dec 14 15:55:25.566: INFO: Deleting pod "simpletest.rc-lhdb4" in namespace "gc-2522" - Dec 14 15:55:25.618: INFO: Deleting pod "simpletest.rc-lhldf" in namespace "gc-2522" - Dec 14 15:55:25.654: INFO: Deleting pod "simpletest.rc-lj5rk" in namespace "gc-2522" - Dec 14 15:55:25.691: INFO: Deleting pod "simpletest.rc-mp5x2" in namespace "gc-2522" - Dec 14 15:55:25.718: INFO: Deleting pod "simpletest.rc-n4ct4" in namespace "gc-2522" - Dec 14 15:55:25.742: INFO: Deleting pod "simpletest.rc-n7lj7" in namespace "gc-2522" - Dec 14 15:55:25.760: INFO: Deleting pod "simpletest.rc-nb4m6" in namespace "gc-2522" - Dec 14 15:55:25.824: INFO: Deleting pod "simpletest.rc-nx4nq" in namespace "gc-2522" - Dec 14 15:55:25.854: INFO: Deleting pod "simpletest.rc-p8jbf" in namespace "gc-2522" - Dec 14 15:55:25.886: INFO: Deleting pod "simpletest.rc-pcqvj" in namespace "gc-2522" - Dec 14 15:55:25.934: INFO: Deleting pod "simpletest.rc-pdgxn" in namespace "gc-2522" - Dec 14 15:55:25.959: INFO: Deleting pod "simpletest.rc-pwqx6" in namespace "gc-2522" - Dec 14 15:55:25.980: INFO: Deleting pod "simpletest.rc-q67qc" in namespace "gc-2522" - Dec 14 15:55:26.026: INFO: Deleting pod "simpletest.rc-qfhg7" in namespace "gc-2522" - Dec 14 15:55:26.068: INFO: Deleting pod "simpletest.rc-qj7lr" in namespace "gc-2522" - Dec 14 15:55:26.097: INFO: Deleting pod "simpletest.rc-qrfkp" in namespace "gc-2522" - Dec 14 15:55:26.131: INFO: Deleting pod "simpletest.rc-rc7vf" in namespace "gc-2522" - Dec 14 15:55:26.176: INFO: Deleting pod "simpletest.rc-rd6xv" in namespace "gc-2522" - Dec 14 15:55:26.210: INFO: Deleting pod "simpletest.rc-rdhdv" in namespace "gc-2522" - Dec 14 15:55:26.268: INFO: Deleting pod "simpletest.rc-sl2mp" in namespace "gc-2522" - Dec 14 15:55:26.303: INFO: Deleting pod "simpletest.rc-tbfzg" in namespace "gc-2522" - Dec 14 15:55:26.345: INFO: Deleting pod "simpletest.rc-td7j2" in namespace "gc-2522" - Dec 14 15:55:26.396: INFO: Deleting pod "simpletest.rc-tlk5d" in namespace "gc-2522" - Dec 14 15:55:26.444: INFO: Deleting pod "simpletest.rc-tmlb9" in namespace "gc-2522" - Dec 14 15:55:26.506: INFO: Deleting pod "simpletest.rc-v2dqg" in namespace "gc-2522" - Dec 14 15:55:26.522: INFO: Deleting pod "simpletest.rc-v9gjf" in namespace "gc-2522" - Dec 14 15:55:26.550: INFO: Deleting pod "simpletest.rc-vgvqx" in namespace "gc-2522" - Dec 14 15:55:26.619: INFO: Deleting pod "simpletest.rc-vzqct" in namespace "gc-2522" - Dec 14 15:55:26.696: INFO: Deleting pod "simpletest.rc-w5hnj" in namespace "gc-2522" - Dec 14 15:55:26.731: INFO: Deleting pod "simpletest.rc-w5tl9" in namespace "gc-2522" - Dec 14 15:55:26.751: INFO: Deleting pod "simpletest.rc-w6m5b" in namespace "gc-2522" - Dec 14 15:55:26.779: INFO: Deleting pod "simpletest.rc-wkfqf" in namespace "gc-2522" - Dec 14 15:55:26.821: INFO: Deleting pod "simpletest.rc-xgh98" in namespace "gc-2522" - Dec 14 15:55:26.854: INFO: Deleting pod "simpletest.rc-xj4dr" in namespace "gc-2522" - Dec 14 15:55:26.880: INFO: Deleting pod "simpletest.rc-xmkk4" in namespace "gc-2522" - Dec 14 15:55:26.923: INFO: Deleting pod "simpletest.rc-xs97t" in namespace "gc-2522" - Dec 14 15:55:26.948: INFO: Deleting pod "simpletest.rc-xsns8" in namespace "gc-2522" - Dec 14 15:55:26.980: INFO: Deleting pod "simpletest.rc-xwk68" in namespace "gc-2522" - Dec 14 15:55:27.020: INFO: Deleting pod "simpletest.rc-z49k8" in namespace "gc-2522" - Dec 14 15:55:27.048: INFO: Deleting pod "simpletest.rc-z5mgj" in namespace "gc-2522" - Dec 14 15:55:27.088: INFO: Deleting pod "simpletest.rc-z99b2" in namespace "gc-2522" - Dec 14 15:55:27.300: INFO: Deleting pod "simpletest.rc-zgmml" in namespace "gc-2522" - [AfterEach] [sig-api-machinery] Garbage collector + STEP: Creating a kubernetes client 02/25/23 09:58:25.74 + Feb 25 09:58:25.741: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 09:58:25.743 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:58:25.825 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:58:25.83 + [BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 + [It] should provide container's memory request [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:234 + STEP: Creating a pod to test downward API volume plugin 02/25/23 09:58:25.836 + Feb 25 09:58:25.859: INFO: Waiting up to 5m0s for pod "downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66" in namespace "projected-9269" to be "Succeeded or Failed" + Feb 25 09:58:25.879: INFO: Pod "downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66": Phase="Pending", Reason="", readiness=false. Elapsed: 19.192158ms + Feb 25 09:58:27.886: INFO: Pod "downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026824685s + Feb 25 09:58:29.888: INFO: Pod "downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.028054387s + STEP: Saw pod success 02/25/23 09:58:29.888 + Feb 25 09:58:29.889: INFO: Pod "downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66" satisfied condition "Succeeded or Failed" + Feb 25 09:58:29.895: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66 container client-container: + STEP: delete the pod 02/25/23 09:58:29.922 + Feb 25 09:58:29.945: INFO: Waiting for pod downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66 to disappear + Feb 25 09:58:29.952: INFO: Pod downwardapi-volume-1dc72d0a-01f6-472a-ae92-2dcb3fd68e66 no longer exists + [AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 - Dec 14 15:55:27.375: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "gc-2522" for this suite. 12/14/22 15:55:27.382 + Feb 25 09:58:29.953: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-9269" for this suite. 02/25/23 09:58:29.961 << End Captured GinkgoWriter Output ------------------------------ -SSSS +SSSSSSSSSSSS ------------------------------ -[sig-network] Networking Granular Checks: Pods - should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:122 -[BeforeEach] [sig-network] Networking +[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] + should list, patch and delete a collection of StatefulSets [Conformance] + test/e2e/apps/statefulset.go:906 +[BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:55:27.406 -Dec 14 15:55:27.406: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename pod-network-test 12/14/22 15:55:27.414 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:55:27.449 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:55:27.457 -[It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:122 -STEP: Performing setup for networking test in namespace pod-network-test-2204 12/14/22 15:55:27.461 -STEP: creating a selector 12/14/22 15:55:27.461 -STEP: Creating the service pods in kubernetes 12/14/22 15:55:27.462 -Dec 14 15:55:27.462: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable -Dec 14 15:55:27.529: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-2204" to be "running and ready" -Dec 14 15:55:27.545: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 16.582345ms -Dec 14 15:55:27.546: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 15:55:29.555: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026431322s -Dec 14 15:55:29.555: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 15:55:31.555: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.025912154s -Dec 14 15:55:31.555: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 15:55:33.552: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.023277202s -Dec 14 15:55:33.553: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 15:55:35.551: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.02248239s -Dec 14 15:55:35.552: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 15:55:37.559: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.030312401s -Dec 14 15:55:37.560: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 15:55:39.556: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 12.02742607s -Dec 14 15:55:39.557: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 15:55:41.558: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 14.028641492s -Dec 14 15:55:41.558: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 15:55:43.555: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 16.025809134s -Dec 14 15:55:43.555: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 15:55:45.556: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 18.027498462s -Dec 14 15:55:45.556: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 15:55:47.555: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 20.026436685s -Dec 14 15:55:47.556: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 15:55:49.555: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 22.025612144s -Dec 14 15:55:49.555: INFO: The phase of Pod netserver-0 is Running (Ready = true) -Dec 14 15:55:49.555: INFO: Pod "netserver-0" satisfied condition "running and ready" -Dec 14 15:55:49.562: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-2204" to be "running and ready" -Dec 14 15:55:49.568: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 6.125989ms -Dec 14 15:55:49.568: INFO: The phase of Pod netserver-1 is Running (Ready = true) -Dec 14 15:55:49.568: INFO: Pod "netserver-1" satisfied condition "running and ready" -Dec 14 15:55:49.573: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-2204" to be "running and ready" -Dec 14 15:55:49.578: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 4.789576ms -Dec 14 15:55:49.578: INFO: The phase of Pod netserver-2 is Running (Ready = true) -Dec 14 15:55:49.578: INFO: Pod "netserver-2" satisfied condition "running and ready" -STEP: Creating test pods 12/14/22 15:55:49.584 -Dec 14 15:55:49.603: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-2204" to be "running" -Dec 14 15:55:49.608: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 4.304776ms -Dec 14 15:55:51.621: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.017911023s -Dec 14 15:55:51.621: INFO: Pod "test-container-pod" satisfied condition "running" -Dec 14 15:55:51.628: INFO: Waiting up to 5m0s for pod "host-test-container-pod" in namespace "pod-network-test-2204" to be "running" -Dec 14 15:55:51.633: INFO: Pod "host-test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 5.624922ms -Dec 14 15:55:51.633: INFO: Pod "host-test-container-pod" satisfied condition "running" -Dec 14 15:55:51.638: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 -Dec 14 15:55:51.638: INFO: Going to poll 10.233.64.46 on port 8081 at least 0 times, with a maximum of 39 tries before failing -Dec 14 15:55:51.643: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.233.64.46 8081 | grep -v '^\s*$'] Namespace:pod-network-test-2204 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 15:55:51.644: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 15:55:51.647: INFO: ExecWithOptions: Clientset creation -Dec 14 15:55:51.647: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-2204/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostName+%7C+nc+-w+1+-u+10.233.64.46+8081+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) -Dec 14 15:55:52.801: INFO: Found all 1 expected endpoints: [netserver-0] -Dec 14 15:55:52.802: INFO: Going to poll 10.233.65.49 on port 8081 at least 0 times, with a maximum of 39 tries before failing -Dec 14 15:55:52.807: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.233.65.49 8081 | grep -v '^\s*$'] Namespace:pod-network-test-2204 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 15:55:52.807: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 15:55:52.809: INFO: ExecWithOptions: Clientset creation -Dec 14 15:55:52.809: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-2204/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostName+%7C+nc+-w+1+-u+10.233.65.49+8081+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) -Dec 14 15:55:53.922: INFO: Found all 1 expected endpoints: [netserver-1] -Dec 14 15:55:53.922: INFO: Going to poll 10.233.66.85 on port 8081 at least 0 times, with a maximum of 39 tries before failing -Dec 14 15:55:53.928: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.233.66.85 8081 | grep -v '^\s*$'] Namespace:pod-network-test-2204 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 15:55:53.928: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 15:55:53.929: INFO: ExecWithOptions: Clientset creation -Dec 14 15:55:53.929: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-2204/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostName+%7C+nc+-w+1+-u+10.233.66.85+8081+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) -Dec 14 15:55:55.036: INFO: Found all 1 expected endpoints: [netserver-2] -[AfterEach] [sig-network] Networking - test/e2e/framework/framework.go:187 -Dec 14 15:55:55.037: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pod-network-test-2204" for this suite. 12/14/22 15:55:55.046 -{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]","completed":38,"skipped":589,"failed":0} ------------------------------- -• [SLOW TEST] [27.653 seconds] -[sig-network] Networking -test/e2e/common/network/framework.go:23 - Granular Checks: Pods - test/e2e/common/network/networking.go:32 - should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:122 +STEP: Creating a kubernetes client 02/25/23 09:58:29.978 +Feb 25 09:58:29.978: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename statefulset 02/25/23 09:58:29.982 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:58:30.013 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:58:30.019 +[BeforeEach] [sig-apps] StatefulSet + test/e2e/apps/statefulset.go:96 +[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:111 +STEP: Creating service test in namespace statefulset-182 02/25/23 09:58:30.023 +[It] should list, patch and delete a collection of StatefulSets [Conformance] + test/e2e/apps/statefulset.go:906 +Feb 25 09:58:30.062: INFO: Found 0 stateful pods, waiting for 1 +Feb 25 09:58:40.070: INFO: Waiting for pod test-ss-0 to enter Running - Ready=true, currently Running - Ready=true +STEP: patching the StatefulSet 02/25/23 09:58:40.08 +W0225 09:58:40.095092 13 warnings.go:70] unknown field "spec.template.spec.TerminationGracePeriodSeconds" +Feb 25 09:58:40.109: INFO: Found 1 stateful pods, waiting for 2 +Feb 25 09:58:50.121: INFO: Waiting for pod test-ss-0 to enter Running - Ready=true, currently Running - Ready=true +Feb 25 09:58:50.121: INFO: Waiting for pod test-ss-1 to enter Running - Ready=true, currently Running - Ready=true +STEP: Listing all StatefulSets 02/25/23 09:58:50.132 +STEP: Delete all of the StatefulSets 02/25/23 09:58:50.136 +STEP: Verify that StatefulSets have been deleted 02/25/23 09:58:50.147 +[AfterEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:122 +Feb 25 09:58:50.154: INFO: Deleting all statefulset in ns statefulset-182 +[AfterEach] [sig-apps] StatefulSet + test/e2e/framework/framework.go:187 +Feb 25 09:58:50.175: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "statefulset-182" for this suite. 02/25/23 09:58:50.184 +{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should list, patch and delete a collection of StatefulSets [Conformance]","completed":39,"skipped":992,"failed":0} +------------------------------ +• [SLOW TEST] [20.230 seconds] +[sig-apps] StatefulSet +test/e2e/apps/framework.go:23 + Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:101 + should list, patch and delete a collection of StatefulSets [Conformance] + test/e2e/apps/statefulset.go:906 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Networking + [BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:55:27.406 - Dec 14 15:55:27.406: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename pod-network-test 12/14/22 15:55:27.414 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:55:27.449 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:55:27.457 - [It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:122 - STEP: Performing setup for networking test in namespace pod-network-test-2204 12/14/22 15:55:27.461 - STEP: creating a selector 12/14/22 15:55:27.461 - STEP: Creating the service pods in kubernetes 12/14/22 15:55:27.462 - Dec 14 15:55:27.462: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable - Dec 14 15:55:27.529: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-2204" to be "running and ready" - Dec 14 15:55:27.545: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 16.582345ms - Dec 14 15:55:27.546: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 15:55:29.555: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026431322s - Dec 14 15:55:29.555: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 15:55:31.555: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.025912154s - Dec 14 15:55:31.555: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 15:55:33.552: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.023277202s - Dec 14 15:55:33.553: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 15:55:35.551: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.02248239s - Dec 14 15:55:35.552: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 15:55:37.559: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.030312401s - Dec 14 15:55:37.560: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 15:55:39.556: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 12.02742607s - Dec 14 15:55:39.557: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 15:55:41.558: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 14.028641492s - Dec 14 15:55:41.558: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 15:55:43.555: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 16.025809134s - Dec 14 15:55:43.555: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 15:55:45.556: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 18.027498462s - Dec 14 15:55:45.556: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 15:55:47.555: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 20.026436685s - Dec 14 15:55:47.556: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 15:55:49.555: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 22.025612144s - Dec 14 15:55:49.555: INFO: The phase of Pod netserver-0 is Running (Ready = true) - Dec 14 15:55:49.555: INFO: Pod "netserver-0" satisfied condition "running and ready" - Dec 14 15:55:49.562: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-2204" to be "running and ready" - Dec 14 15:55:49.568: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 6.125989ms - Dec 14 15:55:49.568: INFO: The phase of Pod netserver-1 is Running (Ready = true) - Dec 14 15:55:49.568: INFO: Pod "netserver-1" satisfied condition "running and ready" - Dec 14 15:55:49.573: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-2204" to be "running and ready" - Dec 14 15:55:49.578: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 4.789576ms - Dec 14 15:55:49.578: INFO: The phase of Pod netserver-2 is Running (Ready = true) - Dec 14 15:55:49.578: INFO: Pod "netserver-2" satisfied condition "running and ready" - STEP: Creating test pods 12/14/22 15:55:49.584 - Dec 14 15:55:49.603: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-2204" to be "running" - Dec 14 15:55:49.608: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 4.304776ms - Dec 14 15:55:51.621: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.017911023s - Dec 14 15:55:51.621: INFO: Pod "test-container-pod" satisfied condition "running" - Dec 14 15:55:51.628: INFO: Waiting up to 5m0s for pod "host-test-container-pod" in namespace "pod-network-test-2204" to be "running" - Dec 14 15:55:51.633: INFO: Pod "host-test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 5.624922ms - Dec 14 15:55:51.633: INFO: Pod "host-test-container-pod" satisfied condition "running" - Dec 14 15:55:51.638: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 - Dec 14 15:55:51.638: INFO: Going to poll 10.233.64.46 on port 8081 at least 0 times, with a maximum of 39 tries before failing - Dec 14 15:55:51.643: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.233.64.46 8081 | grep -v '^\s*$'] Namespace:pod-network-test-2204 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 15:55:51.644: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 15:55:51.647: INFO: ExecWithOptions: Clientset creation - Dec 14 15:55:51.647: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-2204/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostName+%7C+nc+-w+1+-u+10.233.64.46+8081+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) - Dec 14 15:55:52.801: INFO: Found all 1 expected endpoints: [netserver-0] - Dec 14 15:55:52.802: INFO: Going to poll 10.233.65.49 on port 8081 at least 0 times, with a maximum of 39 tries before failing - Dec 14 15:55:52.807: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.233.65.49 8081 | grep -v '^\s*$'] Namespace:pod-network-test-2204 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 15:55:52.807: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 15:55:52.809: INFO: ExecWithOptions: Clientset creation - Dec 14 15:55:52.809: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-2204/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostName+%7C+nc+-w+1+-u+10.233.65.49+8081+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) - Dec 14 15:55:53.922: INFO: Found all 1 expected endpoints: [netserver-1] - Dec 14 15:55:53.922: INFO: Going to poll 10.233.66.85 on port 8081 at least 0 times, with a maximum of 39 tries before failing - Dec 14 15:55:53.928: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.233.66.85 8081 | grep -v '^\s*$'] Namespace:pod-network-test-2204 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 15:55:53.928: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 15:55:53.929: INFO: ExecWithOptions: Clientset creation - Dec 14 15:55:53.929: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-2204/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostName+%7C+nc+-w+1+-u+10.233.66.85+8081+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) - Dec 14 15:55:55.036: INFO: Found all 1 expected endpoints: [netserver-2] - [AfterEach] [sig-network] Networking + STEP: Creating a kubernetes client 02/25/23 09:58:29.978 + Feb 25 09:58:29.978: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename statefulset 02/25/23 09:58:29.982 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:58:30.013 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:58:30.019 + [BeforeEach] [sig-apps] StatefulSet + test/e2e/apps/statefulset.go:96 + [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:111 + STEP: Creating service test in namespace statefulset-182 02/25/23 09:58:30.023 + [It] should list, patch and delete a collection of StatefulSets [Conformance] + test/e2e/apps/statefulset.go:906 + Feb 25 09:58:30.062: INFO: Found 0 stateful pods, waiting for 1 + Feb 25 09:58:40.070: INFO: Waiting for pod test-ss-0 to enter Running - Ready=true, currently Running - Ready=true + STEP: patching the StatefulSet 02/25/23 09:58:40.08 + W0225 09:58:40.095092 13 warnings.go:70] unknown field "spec.template.spec.TerminationGracePeriodSeconds" + Feb 25 09:58:40.109: INFO: Found 1 stateful pods, waiting for 2 + Feb 25 09:58:50.121: INFO: Waiting for pod test-ss-0 to enter Running - Ready=true, currently Running - Ready=true + Feb 25 09:58:50.121: INFO: Waiting for pod test-ss-1 to enter Running - Ready=true, currently Running - Ready=true + STEP: Listing all StatefulSets 02/25/23 09:58:50.132 + STEP: Delete all of the StatefulSets 02/25/23 09:58:50.136 + STEP: Verify that StatefulSets have been deleted 02/25/23 09:58:50.147 + [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:122 + Feb 25 09:58:50.154: INFO: Deleting all statefulset in ns statefulset-182 + [AfterEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:187 - Dec 14 15:55:55.037: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "pod-network-test-2204" for this suite. 12/14/22 15:55:55.046 + Feb 25 09:58:50.175: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "statefulset-182" for this suite. 02/25/23 09:58:50.184 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:176 -[BeforeEach] [sig-storage] EmptyDir volumes +[sig-storage] ConfigMap + should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:46 +[BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:55:55.078 -Dec 14 15:55:55.078: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir 12/14/22 15:55:55.08 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:55:55.111 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:55:55.114 -[It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:176 -STEP: Creating a pod to test emptydir 0666 on node default medium 12/14/22 15:55:55.119 -Dec 14 15:55:55.140: INFO: Waiting up to 5m0s for pod "pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2" in namespace "emptydir-6341" to be "Succeeded or Failed" -Dec 14 15:55:55.158: INFO: Pod "pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2": Phase="Pending", Reason="", readiness=false. Elapsed: 18.027587ms -Dec 14 15:55:57.168: INFO: Pod "pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027925988s -Dec 14 15:55:59.164: INFO: Pod "pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023894221s -STEP: Saw pod success 12/14/22 15:55:59.164 -Dec 14 15:55:59.164: INFO: Pod "pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2" satisfied condition "Succeeded or Failed" -Dec 14 15:55:59.169: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2 container test-container: -STEP: delete the pod 12/14/22 15:55:59.207 -Dec 14 15:55:59.227: INFO: Waiting for pod pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2 to disappear -Dec 14 15:55:59.232: INFO: Pod pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2 no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +STEP: Creating a kubernetes client 02/25/23 09:58:50.213 +Feb 25 09:58:50.213: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename configmap 02/25/23 09:58:50.219 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:58:50.27 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:58:50.275 +[It] should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:46 +STEP: Creating configMap with name configmap-test-volume-5707faaf-4493-4ea8-9bf0-7f18fc12086e 02/25/23 09:58:50.283 +STEP: Creating a pod to test consume configMaps 02/25/23 09:58:50.291 +Feb 25 09:58:50.312: INFO: Waiting up to 5m0s for pod "pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871" in namespace "configmap-1593" to be "Succeeded or Failed" +Feb 25 09:58:50.324: INFO: Pod "pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871": Phase="Pending", Reason="", readiness=false. Elapsed: 12.293056ms +Feb 25 09:58:52.331: INFO: Pod "pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019280995s +Feb 25 09:58:54.336: INFO: Pod "pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023849384s +STEP: Saw pod success 02/25/23 09:58:54.336 +Feb 25 09:58:54.337: INFO: Pod "pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871" satisfied condition "Succeeded or Failed" +Feb 25 09:58:54.345: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871 container agnhost-container: +STEP: delete the pod 02/25/23 09:58:54.358 +Feb 25 09:58:54.397: INFO: Waiting for pod pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871 to disappear +Feb 25 09:58:54.405: INFO: Pod pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871 no longer exists +[AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 -Dec 14 15:55:59.232: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-6341" for this suite. 12/14/22 15:55:59.239 -{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","completed":39,"skipped":658,"failed":0} +Feb 25 09:58:54.406: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-1593" for this suite. 02/25/23 09:58:54.419 +{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance]","completed":40,"skipped":1000,"failed":0} ------------------------------ -• [4.171 seconds] -[sig-storage] EmptyDir volumes +• [4.221 seconds] +[sig-storage] ConfigMap test/e2e/common/storage/framework.go:23 - should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:176 + should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:46 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir volumes + [BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:55:55.078 - Dec 14 15:55:55.078: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir 12/14/22 15:55:55.08 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:55:55.111 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:55:55.114 - [It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:176 - STEP: Creating a pod to test emptydir 0666 on node default medium 12/14/22 15:55:55.119 - Dec 14 15:55:55.140: INFO: Waiting up to 5m0s for pod "pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2" in namespace "emptydir-6341" to be "Succeeded or Failed" - Dec 14 15:55:55.158: INFO: Pod "pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2": Phase="Pending", Reason="", readiness=false. Elapsed: 18.027587ms - Dec 14 15:55:57.168: INFO: Pod "pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027925988s - Dec 14 15:55:59.164: INFO: Pod "pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023894221s - STEP: Saw pod success 12/14/22 15:55:59.164 - Dec 14 15:55:59.164: INFO: Pod "pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2" satisfied condition "Succeeded or Failed" - Dec 14 15:55:59.169: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2 container test-container: - STEP: delete the pod 12/14/22 15:55:59.207 - Dec 14 15:55:59.227: INFO: Waiting for pod pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2 to disappear - Dec 14 15:55:59.232: INFO: Pod pod-b238f4a9-7170-423b-bcd1-a0ec442ec2c2 no longer exists - [AfterEach] [sig-storage] EmptyDir volumes + STEP: Creating a kubernetes client 02/25/23 09:58:50.213 + Feb 25 09:58:50.213: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename configmap 02/25/23 09:58:50.219 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:58:50.27 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:58:50.275 + [It] should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:46 + STEP: Creating configMap with name configmap-test-volume-5707faaf-4493-4ea8-9bf0-7f18fc12086e 02/25/23 09:58:50.283 + STEP: Creating a pod to test consume configMaps 02/25/23 09:58:50.291 + Feb 25 09:58:50.312: INFO: Waiting up to 5m0s for pod "pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871" in namespace "configmap-1593" to be "Succeeded or Failed" + Feb 25 09:58:50.324: INFO: Pod "pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871": Phase="Pending", Reason="", readiness=false. Elapsed: 12.293056ms + Feb 25 09:58:52.331: INFO: Pod "pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019280995s + Feb 25 09:58:54.336: INFO: Pod "pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.023849384s + STEP: Saw pod success 02/25/23 09:58:54.336 + Feb 25 09:58:54.337: INFO: Pod "pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871" satisfied condition "Succeeded or Failed" + Feb 25 09:58:54.345: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871 container agnhost-container: + STEP: delete the pod 02/25/23 09:58:54.358 + Feb 25 09:58:54.397: INFO: Waiting for pod pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871 to disappear + Feb 25 09:58:54.405: INFO: Pod pod-configmaps-847a1d7c-f976-4565-bc50-52e15b459871 no longer exists + [AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 - Dec 14 15:55:59.232: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-6341" for this suite. 12/14/22 15:55:59.239 + Feb 25 09:58:54.406: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "configmap-1593" for this suite. 02/25/23 09:58:54.419 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Downward API - should provide host IP as an env var [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:89 -[BeforeEach] [sig-node] Downward API +[sig-auth] ServiceAccounts + should run through the lifecycle of a ServiceAccount [Conformance] + test/e2e/auth/service_accounts.go:646 +[BeforeEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:55:59.259 -Dec 14 15:55:59.260: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 15:55:59.262 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:55:59.288 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:55:59.292 -[It] should provide host IP as an env var [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:89 -STEP: Creating a pod to test downward api env vars 12/14/22 15:55:59.295 -Dec 14 15:55:59.308: INFO: Waiting up to 5m0s for pod "downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835" in namespace "downward-api-1882" to be "Succeeded or Failed" -Dec 14 15:55:59.318: INFO: Pod "downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835": Phase="Pending", Reason="", readiness=false. Elapsed: 10.064354ms -Dec 14 15:56:01.326: INFO: Pod "downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018806416s -Dec 14 15:56:03.329: INFO: Pod "downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.02136935s -STEP: Saw pod success 12/14/22 15:56:03.329 -Dec 14 15:56:03.329: INFO: Pod "downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835" satisfied condition "Succeeded or Failed" -Dec 14 15:56:03.334: INFO: Trying to get logs from node kiedai4iakae-3 pod downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835 container dapi-container: -STEP: delete the pod 12/14/22 15:56:03.35 -Dec 14 15:56:03.370: INFO: Waiting for pod downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835 to disappear -Dec 14 15:56:03.374: INFO: Pod downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835 no longer exists -[AfterEach] [sig-node] Downward API +STEP: Creating a kubernetes client 02/25/23 09:58:54.441 +Feb 25 09:58:54.441: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename svcaccounts 02/25/23 09:58:54.444 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:58:54.471 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:58:54.477 +[It] should run through the lifecycle of a ServiceAccount [Conformance] + test/e2e/auth/service_accounts.go:646 +STEP: creating a ServiceAccount 02/25/23 09:58:54.482 +STEP: watching for the ServiceAccount to be added 02/25/23 09:58:54.502 +STEP: patching the ServiceAccount 02/25/23 09:58:54.505 +STEP: finding ServiceAccount in list of all ServiceAccounts (by LabelSelector) 02/25/23 09:58:54.515 +STEP: deleting the ServiceAccount 02/25/23 09:58:54.523 +[AfterEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:187 -Dec 14 15:56:03.374: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-1882" for this suite. 12/14/22 15:56:03.38 -{"msg":"PASSED [sig-node] Downward API should provide host IP as an env var [NodeConformance] [Conformance]","completed":40,"skipped":687,"failed":0} +Feb 25 09:58:54.552: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "svcaccounts-4786" for this suite. 02/25/23 09:58:54.561 +{"msg":"PASSED [sig-auth] ServiceAccounts should run through the lifecycle of a ServiceAccount [Conformance]","completed":41,"skipped":1016,"failed":0} ------------------------------ -• [4.129 seconds] -[sig-node] Downward API -test/e2e/common/node/framework.go:23 - should provide host IP as an env var [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:89 +• [0.136 seconds] +[sig-auth] ServiceAccounts +test/e2e/auth/framework.go:23 + should run through the lifecycle of a ServiceAccount [Conformance] + test/e2e/auth/service_accounts.go:646 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Downward API + [BeforeEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:55:59.259 - Dec 14 15:55:59.260: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 15:55:59.262 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:55:59.288 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:55:59.292 - [It] should provide host IP as an env var [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:89 - STEP: Creating a pod to test downward api env vars 12/14/22 15:55:59.295 - Dec 14 15:55:59.308: INFO: Waiting up to 5m0s for pod "downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835" in namespace "downward-api-1882" to be "Succeeded or Failed" - Dec 14 15:55:59.318: INFO: Pod "downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835": Phase="Pending", Reason="", readiness=false. Elapsed: 10.064354ms - Dec 14 15:56:01.326: INFO: Pod "downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018806416s - Dec 14 15:56:03.329: INFO: Pod "downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.02136935s - STEP: Saw pod success 12/14/22 15:56:03.329 - Dec 14 15:56:03.329: INFO: Pod "downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835" satisfied condition "Succeeded or Failed" - Dec 14 15:56:03.334: INFO: Trying to get logs from node kiedai4iakae-3 pod downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835 container dapi-container: - STEP: delete the pod 12/14/22 15:56:03.35 - Dec 14 15:56:03.370: INFO: Waiting for pod downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835 to disappear - Dec 14 15:56:03.374: INFO: Pod downward-api-c115606c-466b-4f85-a0f7-1b06d9e7f835 no longer exists - [AfterEach] [sig-node] Downward API + STEP: Creating a kubernetes client 02/25/23 09:58:54.441 + Feb 25 09:58:54.441: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename svcaccounts 02/25/23 09:58:54.444 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:58:54.471 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:58:54.477 + [It] should run through the lifecycle of a ServiceAccount [Conformance] + test/e2e/auth/service_accounts.go:646 + STEP: creating a ServiceAccount 02/25/23 09:58:54.482 + STEP: watching for the ServiceAccount to be added 02/25/23 09:58:54.502 + STEP: patching the ServiceAccount 02/25/23 09:58:54.505 + STEP: finding ServiceAccount in list of all ServiceAccounts (by LabelSelector) 02/25/23 09:58:54.515 + STEP: deleting the ServiceAccount 02/25/23 09:58:54.523 + [AfterEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:187 - Dec 14 15:56:03.374: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-1882" for this suite. 12/14/22 15:56:03.38 + Feb 25 09:58:54.552: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "svcaccounts-4786" for this suite. 02/25/23 09:58:54.561 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-storage] Projected downwardAPI - should provide container's cpu request [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:220 -[BeforeEach] [sig-storage] Projected downwardAPI +[sig-storage] EmptyDir volumes + should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:106 +[BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:56:03.404 -Dec 14 15:56:03.405: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 15:56:03.407 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:56:03.433 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:56:03.436 -[BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 -[It] should provide container's cpu request [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:220 -STEP: Creating a pod to test downward API volume plugin 12/14/22 15:56:03.44 -Dec 14 15:56:03.457: INFO: Waiting up to 5m0s for pod "downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8" in namespace "projected-4574" to be "Succeeded or Failed" -Dec 14 15:56:03.473: INFO: Pod "downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8": Phase="Pending", Reason="", readiness=false. Elapsed: 15.778796ms -Dec 14 15:56:05.503: INFO: Pod "downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.045818148s -Dec 14 15:56:07.487: INFO: Pod "downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.029294806s -STEP: Saw pod success 12/14/22 15:56:07.487 -Dec 14 15:56:07.487: INFO: Pod "downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8" satisfied condition "Succeeded or Failed" -Dec 14 15:56:07.495: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8 container client-container: -STEP: delete the pod 12/14/22 15:56:07.511 -Dec 14 15:56:07.537: INFO: Waiting for pod downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8 to disappear -Dec 14 15:56:07.542: INFO: Pod downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8 no longer exists -[AfterEach] [sig-storage] Projected downwardAPI +STEP: Creating a kubernetes client 02/25/23 09:58:54.578 +Feb 25 09:58:54.578: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir 02/25/23 09:58:54.58 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:58:54.614 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:58:54.619 +[It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:106 +STEP: Creating a pod to test emptydir 0666 on tmpfs 02/25/23 09:58:54.624 +Feb 25 09:58:54.647: INFO: Waiting up to 5m0s for pod "pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7" in namespace "emptydir-6077" to be "Succeeded or Failed" +Feb 25 09:58:54.656: INFO: Pod "pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7": Phase="Pending", Reason="", readiness=false. Elapsed: 8.670861ms +Feb 25 09:58:56.662: INFO: Pod "pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7": Phase="Running", Reason="", readiness=true. Elapsed: 2.014240005s +Feb 25 09:58:58.662: INFO: Pod "pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7": Phase="Running", Reason="", readiness=false. Elapsed: 4.014852967s +Feb 25 09:59:00.664: INFO: Pod "pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016623946s +STEP: Saw pod success 02/25/23 09:59:00.664 +Feb 25 09:59:00.665: INFO: Pod "pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7" satisfied condition "Succeeded or Failed" +Feb 25 09:59:00.671: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7 container test-container: +STEP: delete the pod 02/25/23 09:59:00.682 +Feb 25 09:59:00.701: INFO: Waiting for pod pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7 to disappear +Feb 25 09:59:00.707: INFO: Pod pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 -Dec 14 15:56:07.542: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-4574" for this suite. 12/14/22 15:56:07.552 -{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance]","completed":41,"skipped":740,"failed":0} +Feb 25 09:59:00.708: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-6077" for this suite. 02/25/23 09:59:00.715 +{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","completed":42,"skipped":1016,"failed":0} ------------------------------ -• [4.160 seconds] -[sig-storage] Projected downwardAPI +• [SLOW TEST] [6.147 seconds] +[sig-storage] EmptyDir volumes test/e2e/common/storage/framework.go:23 - should provide container's cpu request [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:220 + should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:106 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected downwardAPI + [BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:56:03.404 - Dec 14 15:56:03.405: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 15:56:03.407 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:56:03.433 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:56:03.436 - [BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 - [It] should provide container's cpu request [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:220 - STEP: Creating a pod to test downward API volume plugin 12/14/22 15:56:03.44 - Dec 14 15:56:03.457: INFO: Waiting up to 5m0s for pod "downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8" in namespace "projected-4574" to be "Succeeded or Failed" - Dec 14 15:56:03.473: INFO: Pod "downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8": Phase="Pending", Reason="", readiness=false. Elapsed: 15.778796ms - Dec 14 15:56:05.503: INFO: Pod "downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.045818148s - Dec 14 15:56:07.487: INFO: Pod "downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.029294806s - STEP: Saw pod success 12/14/22 15:56:07.487 - Dec 14 15:56:07.487: INFO: Pod "downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8" satisfied condition "Succeeded or Failed" - Dec 14 15:56:07.495: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8 container client-container: - STEP: delete the pod 12/14/22 15:56:07.511 - Dec 14 15:56:07.537: INFO: Waiting for pod downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8 to disappear - Dec 14 15:56:07.542: INFO: Pod downwardapi-volume-7512fb0a-2372-4079-b80f-57eba6ed87f8 no longer exists - [AfterEach] [sig-storage] Projected downwardAPI + STEP: Creating a kubernetes client 02/25/23 09:58:54.578 + Feb 25 09:58:54.578: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir 02/25/23 09:58:54.58 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:58:54.614 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:58:54.619 + [It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:106 + STEP: Creating a pod to test emptydir 0666 on tmpfs 02/25/23 09:58:54.624 + Feb 25 09:58:54.647: INFO: Waiting up to 5m0s for pod "pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7" in namespace "emptydir-6077" to be "Succeeded or Failed" + Feb 25 09:58:54.656: INFO: Pod "pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7": Phase="Pending", Reason="", readiness=false. Elapsed: 8.670861ms + Feb 25 09:58:56.662: INFO: Pod "pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7": Phase="Running", Reason="", readiness=true. Elapsed: 2.014240005s + Feb 25 09:58:58.662: INFO: Pod "pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7": Phase="Running", Reason="", readiness=false. Elapsed: 4.014852967s + Feb 25 09:59:00.664: INFO: Pod "pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016623946s + STEP: Saw pod success 02/25/23 09:59:00.664 + Feb 25 09:59:00.665: INFO: Pod "pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7" satisfied condition "Succeeded or Failed" + Feb 25 09:59:00.671: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7 container test-container: + STEP: delete the pod 02/25/23 09:59:00.682 + Feb 25 09:59:00.701: INFO: Waiting for pod pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7 to disappear + Feb 25 09:59:00.707: INFO: Pod pod-c319ded7-fc84-4a55-a5da-c0c40f2934b7 no longer exists + [AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 - Dec 14 15:56:07.542: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-4574" for this suite. 12/14/22 15:56:07.552 + Feb 25 09:59:00.708: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-6077" for this suite. 02/25/23 09:59:00.715 << End Captured GinkgoWriter Output ------------------------------ -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - works for CRD preserving unknown fields in an embedded object [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:235 -[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +SSSSSSSS +------------------------------ +[sig-storage] Downward API volume + should provide container's cpu limit [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:192 +[BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:56:07.566 -Dec 14 15:56:07.566: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 15:56:07.57 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:56:07.6 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:56:07.603 -[It] works for CRD preserving unknown fields in an embedded object [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:235 -Dec 14 15:56:07.610: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: kubectl validation (kubectl create and apply) allows request with any unknown properties 12/14/22 15:56:11.009 -Dec 14 15:56:11.010: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2002 --namespace=crd-publish-openapi-2002 create -f -' -Dec 14 15:56:12.281: INFO: stderr: "" -Dec 14 15:56:12.281: INFO: stdout: "e2e-test-crd-publish-openapi-2502-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n" -Dec 14 15:56:12.281: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2002 --namespace=crd-publish-openapi-2002 delete e2e-test-crd-publish-openapi-2502-crds test-cr' -Dec 14 15:56:12.416: INFO: stderr: "" -Dec 14 15:56:12.417: INFO: stdout: "e2e-test-crd-publish-openapi-2502-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n" -Dec 14 15:56:12.417: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2002 --namespace=crd-publish-openapi-2002 apply -f -' -Dec 14 15:56:12.807: INFO: stderr: "" -Dec 14 15:56:12.807: INFO: stdout: "e2e-test-crd-publish-openapi-2502-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n" -Dec 14 15:56:12.807: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2002 --namespace=crd-publish-openapi-2002 delete e2e-test-crd-publish-openapi-2502-crds test-cr' -Dec 14 15:56:13.037: INFO: stderr: "" -Dec 14 15:56:13.037: INFO: stdout: "e2e-test-crd-publish-openapi-2502-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n" -STEP: kubectl explain works to explain CR 12/14/22 15:56:13.037 -Dec 14 15:56:13.037: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2002 explain e2e-test-crd-publish-openapi-2502-crds' -Dec 14 15:56:14.146: INFO: stderr: "" -Dec 14 15:56:14.146: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-2502-crd\nVERSION: crd-publish-openapi-test-unknown-in-nested.example.com/v1\n\nDESCRIPTION:\n preserve-unknown-properties in nested field for Testing\n\nFIELDS:\n apiVersion\t\n APIVersion defines the versioned schema of this representation of an\n object. Servers should convert recognized schemas to the latest internal\n value, and may reject unrecognized values. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n kind\t\n Kind is a string value representing the REST resource this object\n represents. Servers may infer this from the endpoint the client submits\n requests to. Cannot be updated. In CamelCase. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n metadata\t\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n spec\t<>\n Specification of Waldo\n\n status\t\n Status of Waldo\n\n" -[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 09:59:00.728 +Feb 25 09:59:00.728: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 09:59:00.73 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:59:00.758 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:59:00.762 +[BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 +[It] should provide container's cpu limit [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:192 +STEP: Creating a pod to test downward API volume plugin 02/25/23 09:59:00.766 +Feb 25 09:59:00.781: INFO: Waiting up to 5m0s for pod "downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4" in namespace "downward-api-4638" to be "Succeeded or Failed" +Feb 25 09:59:00.793: INFO: Pod "downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4": Phase="Pending", Reason="", readiness=false. Elapsed: 12.458689ms +Feb 25 09:59:02.802: INFO: Pod "downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021702696s +Feb 25 09:59:04.801: INFO: Pod "downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4": Phase="Pending", Reason="", readiness=false. Elapsed: 4.020572926s +Feb 25 09:59:06.800: INFO: Pod "downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.019859292s +STEP: Saw pod success 02/25/23 09:59:06.801 +Feb 25 09:59:06.801: INFO: Pod "downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4" satisfied condition "Succeeded or Failed" +Feb 25 09:59:06.809: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4 container client-container: +STEP: delete the pod 02/25/23 09:59:06.819 +Feb 25 09:59:06.841: INFO: Waiting for pod downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4 to disappear +Feb 25 09:59:06.849: INFO: Pod downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4 no longer exists +[AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 -Dec 14 15:56:19.522: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-publish-openapi-2002" for this suite. 12/14/22 15:56:19.539 -{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields in an embedded object [Conformance]","completed":42,"skipped":740,"failed":0} +Feb 25 09:59:06.850: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-4638" for this suite. 02/25/23 09:59:06.859 +{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance]","completed":43,"skipped":1024,"failed":0} ------------------------------ -• [SLOW TEST] [11.983 seconds] -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - works for CRD preserving unknown fields in an embedded object [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:235 +• [SLOW TEST] [6.149 seconds] +[sig-storage] Downward API volume +test/e2e/common/storage/framework.go:23 + should provide container's cpu limit [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:192 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + [BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:56:07.566 - Dec 14 15:56:07.566: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 15:56:07.57 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:56:07.6 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:56:07.603 - [It] works for CRD preserving unknown fields in an embedded object [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:235 - Dec 14 15:56:07.610: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: kubectl validation (kubectl create and apply) allows request with any unknown properties 12/14/22 15:56:11.009 - Dec 14 15:56:11.010: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2002 --namespace=crd-publish-openapi-2002 create -f -' - Dec 14 15:56:12.281: INFO: stderr: "" - Dec 14 15:56:12.281: INFO: stdout: "e2e-test-crd-publish-openapi-2502-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n" - Dec 14 15:56:12.281: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2002 --namespace=crd-publish-openapi-2002 delete e2e-test-crd-publish-openapi-2502-crds test-cr' - Dec 14 15:56:12.416: INFO: stderr: "" - Dec 14 15:56:12.417: INFO: stdout: "e2e-test-crd-publish-openapi-2502-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n" - Dec 14 15:56:12.417: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2002 --namespace=crd-publish-openapi-2002 apply -f -' - Dec 14 15:56:12.807: INFO: stderr: "" - Dec 14 15:56:12.807: INFO: stdout: "e2e-test-crd-publish-openapi-2502-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n" - Dec 14 15:56:12.807: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2002 --namespace=crd-publish-openapi-2002 delete e2e-test-crd-publish-openapi-2502-crds test-cr' - Dec 14 15:56:13.037: INFO: stderr: "" - Dec 14 15:56:13.037: INFO: stdout: "e2e-test-crd-publish-openapi-2502-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n" - STEP: kubectl explain works to explain CR 12/14/22 15:56:13.037 - Dec 14 15:56:13.037: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2002 explain e2e-test-crd-publish-openapi-2502-crds' - Dec 14 15:56:14.146: INFO: stderr: "" - Dec 14 15:56:14.146: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-2502-crd\nVERSION: crd-publish-openapi-test-unknown-in-nested.example.com/v1\n\nDESCRIPTION:\n preserve-unknown-properties in nested field for Testing\n\nFIELDS:\n apiVersion\t\n APIVersion defines the versioned schema of this representation of an\n object. Servers should convert recognized schemas to the latest internal\n value, and may reject unrecognized values. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n kind\t\n Kind is a string value representing the REST resource this object\n represents. Servers may infer this from the endpoint the client submits\n requests to. Cannot be updated. In CamelCase. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n metadata\t\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n spec\t<>\n Specification of Waldo\n\n status\t\n Status of Waldo\n\n" - [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 09:59:00.728 + Feb 25 09:59:00.728: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 09:59:00.73 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:59:00.758 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:59:00.762 + [BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 + [It] should provide container's cpu limit [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:192 + STEP: Creating a pod to test downward API volume plugin 02/25/23 09:59:00.766 + Feb 25 09:59:00.781: INFO: Waiting up to 5m0s for pod "downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4" in namespace "downward-api-4638" to be "Succeeded or Failed" + Feb 25 09:59:00.793: INFO: Pod "downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4": Phase="Pending", Reason="", readiness=false. Elapsed: 12.458689ms + Feb 25 09:59:02.802: INFO: Pod "downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021702696s + Feb 25 09:59:04.801: INFO: Pod "downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4": Phase="Pending", Reason="", readiness=false. Elapsed: 4.020572926s + Feb 25 09:59:06.800: INFO: Pod "downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.019859292s + STEP: Saw pod success 02/25/23 09:59:06.801 + Feb 25 09:59:06.801: INFO: Pod "downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4" satisfied condition "Succeeded or Failed" + Feb 25 09:59:06.809: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4 container client-container: + STEP: delete the pod 02/25/23 09:59:06.819 + Feb 25 09:59:06.841: INFO: Waiting for pod downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4 to disappear + Feb 25 09:59:06.849: INFO: Pod downwardapi-volume-529bb24e-9c87-4900-a22b-608ff5530de4 no longer exists + [AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 - Dec 14 15:56:19.522: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "crd-publish-openapi-2002" for this suite. 12/14/22 15:56:19.539 + Feb 25 09:59:06.850: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-4638" for this suite. 02/25/23 09:59:06.859 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSS +SSSS ------------------------------ -[sig-node] ConfigMap - should run through a ConfigMap lifecycle [Conformance] - test/e2e/common/node/configmap.go:168 -[BeforeEach] [sig-node] ConfigMap +[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook + should execute poststart http hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:130 +[BeforeEach] [sig-node] Container Lifecycle Hook test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:56:19.558 -Dec 14 15:56:19.559: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename configmap 12/14/22 15:56:19.56 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:56:19.582 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:56:19.587 -[It] should run through a ConfigMap lifecycle [Conformance] - test/e2e/common/node/configmap.go:168 -STEP: creating a ConfigMap 12/14/22 15:56:19.59 -STEP: fetching the ConfigMap 12/14/22 15:56:19.596 -STEP: patching the ConfigMap 12/14/22 15:56:19.601 -STEP: listing all ConfigMaps in all namespaces with a label selector 12/14/22 15:56:19.608 -STEP: deleting the ConfigMap by collection with a label selector 12/14/22 15:56:19.613 -STEP: listing all ConfigMaps in test namespace 12/14/22 15:56:19.624 -[AfterEach] [sig-node] ConfigMap +STEP: Creating a kubernetes client 02/25/23 09:59:06.879 +Feb 25 09:59:06.880: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-lifecycle-hook 02/25/23 09:59:06.888 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:59:06.918 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:59:06.924 +[BeforeEach] when create a pod with lifecycle hook + test/e2e/common/node/lifecycle_hook.go:55 +STEP: create the container to handle the HTTPGet hook request. 02/25/23 09:59:06.935 +Feb 25 09:59:06.949: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-4641" to be "running and ready" +Feb 25 09:59:06.960: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 10.654973ms +Feb 25 09:59:06.960: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) +Feb 25 09:59:08.969: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 2.019737675s +Feb 25 09:59:08.969: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) +Feb 25 09:59:08.969: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" +[It] should execute poststart http hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:130 +STEP: create the pod with lifecycle hook 02/25/23 09:59:08.981 +Feb 25 09:59:08.994: INFO: Waiting up to 5m0s for pod "pod-with-poststart-http-hook" in namespace "container-lifecycle-hook-4641" to be "running and ready" +Feb 25 09:59:09.002: INFO: Pod "pod-with-poststart-http-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 8.425909ms +Feb 25 09:59:09.002: INFO: The phase of Pod pod-with-poststart-http-hook is Pending, waiting for it to be Running (with Ready = true) +Feb 25 09:59:11.009: INFO: Pod "pod-with-poststart-http-hook": Phase="Running", Reason="", readiness=true. Elapsed: 2.015541172s +Feb 25 09:59:11.009: INFO: The phase of Pod pod-with-poststart-http-hook is Running (Ready = true) +Feb 25 09:59:11.009: INFO: Pod "pod-with-poststart-http-hook" satisfied condition "running and ready" +STEP: check poststart hook 02/25/23 09:59:11.015 +STEP: delete the pod with lifecycle hook 02/25/23 09:59:11.043 +Feb 25 09:59:11.057: INFO: Waiting for pod pod-with-poststart-http-hook to disappear +Feb 25 09:59:11.065: INFO: Pod pod-with-poststart-http-hook still exists +Feb 25 09:59:13.066: INFO: Waiting for pod pod-with-poststart-http-hook to disappear +Feb 25 09:59:13.075: INFO: Pod pod-with-poststart-http-hook still exists +Feb 25 09:59:15.067: INFO: Waiting for pod pod-with-poststart-http-hook to disappear +Feb 25 09:59:15.075: INFO: Pod pod-with-poststart-http-hook no longer exists +[AfterEach] [sig-node] Container Lifecycle Hook test/e2e/framework/framework.go:187 -Dec 14 15:56:19.628: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-1346" for this suite. 12/14/22 15:56:19.632 -{"msg":"PASSED [sig-node] ConfigMap should run through a ConfigMap lifecycle [Conformance]","completed":43,"skipped":755,"failed":0} +Feb 25 09:59:15.075: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-lifecycle-hook-4641" for this suite. 02/25/23 09:59:15.082 +{"msg":"PASSED [sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance]","completed":44,"skipped":1028,"failed":0} ------------------------------ -• [0.083 seconds] -[sig-node] ConfigMap +• [SLOW TEST] [8.215 seconds] +[sig-node] Container Lifecycle Hook test/e2e/common/node/framework.go:23 - should run through a ConfigMap lifecycle [Conformance] - test/e2e/common/node/configmap.go:168 + when create a pod with lifecycle hook + test/e2e/common/node/lifecycle_hook.go:46 + should execute poststart http hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:130 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] ConfigMap + [BeforeEach] [sig-node] Container Lifecycle Hook test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:56:19.558 - Dec 14 15:56:19.559: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename configmap 12/14/22 15:56:19.56 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:56:19.582 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:56:19.587 - [It] should run through a ConfigMap lifecycle [Conformance] - test/e2e/common/node/configmap.go:168 - STEP: creating a ConfigMap 12/14/22 15:56:19.59 - STEP: fetching the ConfigMap 12/14/22 15:56:19.596 - STEP: patching the ConfigMap 12/14/22 15:56:19.601 - STEP: listing all ConfigMaps in all namespaces with a label selector 12/14/22 15:56:19.608 - STEP: deleting the ConfigMap by collection with a label selector 12/14/22 15:56:19.613 - STEP: listing all ConfigMaps in test namespace 12/14/22 15:56:19.624 - [AfterEach] [sig-node] ConfigMap + STEP: Creating a kubernetes client 02/25/23 09:59:06.879 + Feb 25 09:59:06.880: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-lifecycle-hook 02/25/23 09:59:06.888 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:59:06.918 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:59:06.924 + [BeforeEach] when create a pod with lifecycle hook + test/e2e/common/node/lifecycle_hook.go:55 + STEP: create the container to handle the HTTPGet hook request. 02/25/23 09:59:06.935 + Feb 25 09:59:06.949: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-4641" to be "running and ready" + Feb 25 09:59:06.960: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 10.654973ms + Feb 25 09:59:06.960: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) + Feb 25 09:59:08.969: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 2.019737675s + Feb 25 09:59:08.969: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) + Feb 25 09:59:08.969: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" + [It] should execute poststart http hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:130 + STEP: create the pod with lifecycle hook 02/25/23 09:59:08.981 + Feb 25 09:59:08.994: INFO: Waiting up to 5m0s for pod "pod-with-poststart-http-hook" in namespace "container-lifecycle-hook-4641" to be "running and ready" + Feb 25 09:59:09.002: INFO: Pod "pod-with-poststart-http-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 8.425909ms + Feb 25 09:59:09.002: INFO: The phase of Pod pod-with-poststart-http-hook is Pending, waiting for it to be Running (with Ready = true) + Feb 25 09:59:11.009: INFO: Pod "pod-with-poststart-http-hook": Phase="Running", Reason="", readiness=true. Elapsed: 2.015541172s + Feb 25 09:59:11.009: INFO: The phase of Pod pod-with-poststart-http-hook is Running (Ready = true) + Feb 25 09:59:11.009: INFO: Pod "pod-with-poststart-http-hook" satisfied condition "running and ready" + STEP: check poststart hook 02/25/23 09:59:11.015 + STEP: delete the pod with lifecycle hook 02/25/23 09:59:11.043 + Feb 25 09:59:11.057: INFO: Waiting for pod pod-with-poststart-http-hook to disappear + Feb 25 09:59:11.065: INFO: Pod pod-with-poststart-http-hook still exists + Feb 25 09:59:13.066: INFO: Waiting for pod pod-with-poststart-http-hook to disappear + Feb 25 09:59:13.075: INFO: Pod pod-with-poststart-http-hook still exists + Feb 25 09:59:15.067: INFO: Waiting for pod pod-with-poststart-http-hook to disappear + Feb 25 09:59:15.075: INFO: Pod pod-with-poststart-http-hook no longer exists + [AfterEach] [sig-node] Container Lifecycle Hook test/e2e/framework/framework.go:187 - Dec 14 15:56:19.628: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "configmap-1346" for this suite. 12/14/22 15:56:19.632 + Feb 25 09:59:15.075: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-lifecycle-hook-4641" for this suite. 02/25/23 09:59:15.082 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] - Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] - test/e2e/apps/statefulset.go:695 -[BeforeEach] [sig-apps] StatefulSet +[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook + should execute poststart exec hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:97 +[BeforeEach] [sig-node] Container Lifecycle Hook test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:56:19.647 -Dec 14 15:56:19.648: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename statefulset 12/14/22 15:56:19.649 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:56:19.671 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:56:19.675 -[BeforeEach] [sig-apps] StatefulSet - test/e2e/apps/statefulset.go:96 -[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:111 -STEP: Creating service test in namespace statefulset-6999 12/14/22 15:56:19.677 -[It] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] - test/e2e/apps/statefulset.go:695 -STEP: Creating stateful set ss in namespace statefulset-6999 12/14/22 15:56:19.686 -STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-6999 12/14/22 15:56:19.701 -Dec 14 15:56:19.707: INFO: Found 0 stateful pods, waiting for 1 -Dec 14 15:56:29.721: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true -STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod 12/14/22 15:56:29.721 -Dec 14 15:56:29.729: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6999 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' -Dec 14 15:56:30.156: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" -Dec 14 15:56:30.156: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" -Dec 14 15:56:30.157: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - -Dec 14 15:56:30.171: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true -Dec 14 15:56:40.182: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false -Dec 14 15:56:40.182: INFO: Waiting for statefulset status.replicas updated to 0 -Dec 14 15:56:40.212: INFO: POD NODE PHASE GRACE CONDITIONS -Dec 14 15:56:40.212: INFO: ss-0 kiedai4iakae-3 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:19 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:30 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:30 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:19 +0000 UTC }] -Dec 14 15:56:40.212: INFO: -Dec 14 15:56:40.213: INFO: StatefulSet ss has not reached scale 3, at 1 -Dec 14 15:56:41.229: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.991270685s -Dec 14 15:56:42.237: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.975189945s -Dec 14 15:56:43.246: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.966808694s -Dec 14 15:56:44.258: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.95701363s -Dec 14 15:56:45.267: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.945696632s -Dec 14 15:56:46.277: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.936744895s -Dec 14 15:56:47.291: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.926384986s -Dec 14 15:56:48.302: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.912332361s -Dec 14 15:56:49.311: INFO: Verifying statefulset ss doesn't scale past 3 for another 901.930002ms -STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-6999 12/14/22 15:56:50.311 -Dec 14 15:56:50.322: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6999 exec ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Dec 14 15:56:50.573: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" -Dec 14 15:56:50.573: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" -Dec 14 15:56:50.573: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - -Dec 14 15:56:50.574: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6999 exec ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Dec 14 15:56:50.798: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n" -Dec 14 15:56:50.798: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" -Dec 14 15:56:50.798: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - -Dec 14 15:56:50.799: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6999 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Dec 14 15:56:51.039: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n" -Dec 14 15:56:51.039: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" -Dec 14 15:56:51.039: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - -Dec 14 15:56:51.046: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=false -Dec 14 15:57:01.064: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true -Dec 14 15:57:01.064: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true -Dec 14 15:57:01.064: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true -STEP: Scale down will not halt with unhealthy stateful pod 12/14/22 15:57:01.064 -Dec 14 15:57:01.073: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6999 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' -Dec 14 15:57:01.341: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" -Dec 14 15:57:01.341: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" -Dec 14 15:57:01.341: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - -Dec 14 15:57:01.342: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6999 exec ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' -Dec 14 15:57:01.590: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" -Dec 14 15:57:01.590: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" -Dec 14 15:57:01.590: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - -Dec 14 15:57:01.590: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6999 exec ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' -Dec 14 15:57:01.829: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" -Dec 14 15:57:01.829: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" -Dec 14 15:57:01.829: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - -Dec 14 15:57:01.829: INFO: Waiting for statefulset status.replicas updated to 0 -Dec 14 15:57:01.835: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 2 -Dec 14 15:57:11.852: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false -Dec 14 15:57:11.852: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false -Dec 14 15:57:11.853: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false -Dec 14 15:57:11.874: INFO: POD NODE PHASE GRACE CONDITIONS -Dec 14 15:57:11.874: INFO: ss-0 kiedai4iakae-3 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:19 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:19 +0000 UTC }] -Dec 14 15:57:11.874: INFO: ss-1 kiedai4iakae-1 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:40 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:40 +0000 UTC }] -Dec 14 15:57:11.875: INFO: ss-2 kiedai4iakae-2 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:40 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:40 +0000 UTC }] -Dec 14 15:57:11.875: INFO: -Dec 14 15:57:11.875: INFO: StatefulSet ss has not reached scale 0, at 3 -Dec 14 15:57:12.887: INFO: POD NODE PHASE GRACE CONDITIONS -Dec 14 15:57:12.887: INFO: ss-0 kiedai4iakae-3 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:19 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:19 +0000 UTC }] -Dec 14 15:57:12.888: INFO: ss-2 kiedai4iakae-2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:40 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:40 +0000 UTC }] -Dec 14 15:57:12.888: INFO: -Dec 14 15:57:12.888: INFO: StatefulSet ss has not reached scale 0, at 2 -Dec 14 15:57:13.894: INFO: Verifying statefulset ss doesn't scale past 0 for another 7.979160546s -Dec 14 15:57:14.904: INFO: Verifying statefulset ss doesn't scale past 0 for another 6.972597377s -Dec 14 15:57:15.915: INFO: Verifying statefulset ss doesn't scale past 0 for another 5.962082548s -Dec 14 15:57:16.929: INFO: Verifying statefulset ss doesn't scale past 0 for another 4.951435285s -Dec 14 15:57:17.935: INFO: Verifying statefulset ss doesn't scale past 0 for another 3.938553521s -Dec 14 15:57:18.942: INFO: Verifying statefulset ss doesn't scale past 0 for another 2.931382197s -Dec 14 15:57:19.948: INFO: Verifying statefulset ss doesn't scale past 0 for another 1.92418479s -Dec 14 15:57:20.958: INFO: Verifying statefulset ss doesn't scale past 0 for another 919.133634ms -STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-6999 12/14/22 15:57:21.959 -Dec 14 15:57:21.967: INFO: Scaling statefulset ss to 0 -Dec 14 15:57:22.000: INFO: Waiting for statefulset status.replicas updated to 0 -[AfterEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:122 -Dec 14 15:57:22.006: INFO: Deleting all statefulset in ns statefulset-6999 -Dec 14 15:57:22.012: INFO: Scaling statefulset ss to 0 -Dec 14 15:57:22.040: INFO: Waiting for statefulset status.replicas updated to 0 -Dec 14 15:57:22.044: INFO: Deleting statefulset ss -[AfterEach] [sig-apps] StatefulSet +STEP: Creating a kubernetes client 02/25/23 09:59:15.104 +Feb 25 09:59:15.104: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-lifecycle-hook 02/25/23 09:59:15.106 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:59:15.134 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:59:15.138 +[BeforeEach] when create a pod with lifecycle hook + test/e2e/common/node/lifecycle_hook.go:55 +STEP: create the container to handle the HTTPGet hook request. 02/25/23 09:59:15.154 +Feb 25 09:59:15.169: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-2342" to be "running and ready" +Feb 25 09:59:15.185: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 15.236413ms +Feb 25 09:59:15.185: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) +Feb 25 09:59:17.195: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02569601s +Feb 25 09:59:17.195: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) +Feb 25 09:59:19.192: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 4.022451563s +Feb 25 09:59:19.192: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) +Feb 25 09:59:19.192: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" +[It] should execute poststart exec hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:97 +STEP: create the pod with lifecycle hook 02/25/23 09:59:19.198 +Feb 25 09:59:19.209: INFO: Waiting up to 5m0s for pod "pod-with-poststart-exec-hook" in namespace "container-lifecycle-hook-2342" to be "running and ready" +Feb 25 09:59:19.216: INFO: Pod "pod-with-poststart-exec-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 6.597926ms +Feb 25 09:59:19.216: INFO: The phase of Pod pod-with-poststart-exec-hook is Pending, waiting for it to be Running (with Ready = true) +Feb 25 09:59:21.226: INFO: Pod "pod-with-poststart-exec-hook": Phase="Running", Reason="", readiness=true. Elapsed: 2.016274317s +Feb 25 09:59:21.226: INFO: The phase of Pod pod-with-poststart-exec-hook is Running (Ready = true) +Feb 25 09:59:21.226: INFO: Pod "pod-with-poststart-exec-hook" satisfied condition "running and ready" +STEP: check poststart hook 02/25/23 09:59:21.232 +STEP: delete the pod with lifecycle hook 02/25/23 09:59:21.244 +Feb 25 09:59:21.257: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear +Feb 25 09:59:21.263: INFO: Pod pod-with-poststart-exec-hook still exists +Feb 25 09:59:23.264: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear +Feb 25 09:59:23.272: INFO: Pod pod-with-poststart-exec-hook still exists +Feb 25 09:59:25.264: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear +Feb 25 09:59:25.270: INFO: Pod pod-with-poststart-exec-hook no longer exists +[AfterEach] [sig-node] Container Lifecycle Hook test/e2e/framework/framework.go:187 -Dec 14 15:57:22.065: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "statefulset-6999" for this suite. 12/14/22 15:57:22.078 -{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance]","completed":44,"skipped":770,"failed":0} +Feb 25 09:59:25.270: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-lifecycle-hook-2342" for this suite. 02/25/23 09:59:25.278 +{"msg":"PASSED [sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart exec hook properly [NodeConformance] [Conformance]","completed":45,"skipped":1068,"failed":0} ------------------------------ -• [SLOW TEST] [62.448 seconds] -[sig-apps] StatefulSet -test/e2e/apps/framework.go:23 - Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:101 - Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] - test/e2e/apps/statefulset.go:695 +• [SLOW TEST] [10.184 seconds] +[sig-node] Container Lifecycle Hook +test/e2e/common/node/framework.go:23 + when create a pod with lifecycle hook + test/e2e/common/node/lifecycle_hook.go:46 + should execute poststart exec hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:97 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] StatefulSet + [BeforeEach] [sig-node] Container Lifecycle Hook test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:56:19.647 - Dec 14 15:56:19.648: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename statefulset 12/14/22 15:56:19.649 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:56:19.671 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:56:19.675 - [BeforeEach] [sig-apps] StatefulSet - test/e2e/apps/statefulset.go:96 - [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:111 - STEP: Creating service test in namespace statefulset-6999 12/14/22 15:56:19.677 - [It] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] - test/e2e/apps/statefulset.go:695 - STEP: Creating stateful set ss in namespace statefulset-6999 12/14/22 15:56:19.686 - STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-6999 12/14/22 15:56:19.701 - Dec 14 15:56:19.707: INFO: Found 0 stateful pods, waiting for 1 - Dec 14 15:56:29.721: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true - STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod 12/14/22 15:56:29.721 - Dec 14 15:56:29.729: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6999 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' - Dec 14 15:56:30.156: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" - Dec 14 15:56:30.156: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" - Dec 14 15:56:30.157: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - - Dec 14 15:56:30.171: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true - Dec 14 15:56:40.182: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false - Dec 14 15:56:40.182: INFO: Waiting for statefulset status.replicas updated to 0 - Dec 14 15:56:40.212: INFO: POD NODE PHASE GRACE CONDITIONS - Dec 14 15:56:40.212: INFO: ss-0 kiedai4iakae-3 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:19 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:30 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:30 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:19 +0000 UTC }] - Dec 14 15:56:40.212: INFO: - Dec 14 15:56:40.213: INFO: StatefulSet ss has not reached scale 3, at 1 - Dec 14 15:56:41.229: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.991270685s - Dec 14 15:56:42.237: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.975189945s - Dec 14 15:56:43.246: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.966808694s - Dec 14 15:56:44.258: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.95701363s - Dec 14 15:56:45.267: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.945696632s - Dec 14 15:56:46.277: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.936744895s - Dec 14 15:56:47.291: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.926384986s - Dec 14 15:56:48.302: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.912332361s - Dec 14 15:56:49.311: INFO: Verifying statefulset ss doesn't scale past 3 for another 901.930002ms - STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-6999 12/14/22 15:56:50.311 - Dec 14 15:56:50.322: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6999 exec ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' - Dec 14 15:56:50.573: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" - Dec 14 15:56:50.573: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" - Dec 14 15:56:50.573: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - - Dec 14 15:56:50.574: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6999 exec ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' - Dec 14 15:56:50.798: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n" - Dec 14 15:56:50.798: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" - Dec 14 15:56:50.798: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - - Dec 14 15:56:50.799: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6999 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' - Dec 14 15:56:51.039: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n" - Dec 14 15:56:51.039: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" - Dec 14 15:56:51.039: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - - Dec 14 15:56:51.046: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=false - Dec 14 15:57:01.064: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true - Dec 14 15:57:01.064: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true - Dec 14 15:57:01.064: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true - STEP: Scale down will not halt with unhealthy stateful pod 12/14/22 15:57:01.064 - Dec 14 15:57:01.073: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6999 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' - Dec 14 15:57:01.341: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" - Dec 14 15:57:01.341: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" - Dec 14 15:57:01.341: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - - Dec 14 15:57:01.342: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6999 exec ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' - Dec 14 15:57:01.590: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" - Dec 14 15:57:01.590: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" - Dec 14 15:57:01.590: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - - Dec 14 15:57:01.590: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-6999 exec ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' - Dec 14 15:57:01.829: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" - Dec 14 15:57:01.829: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" - Dec 14 15:57:01.829: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - - Dec 14 15:57:01.829: INFO: Waiting for statefulset status.replicas updated to 0 - Dec 14 15:57:01.835: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 2 - Dec 14 15:57:11.852: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false - Dec 14 15:57:11.852: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false - Dec 14 15:57:11.853: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false - Dec 14 15:57:11.874: INFO: POD NODE PHASE GRACE CONDITIONS - Dec 14 15:57:11.874: INFO: ss-0 kiedai4iakae-3 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:19 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:19 +0000 UTC }] - Dec 14 15:57:11.874: INFO: ss-1 kiedai4iakae-1 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:40 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:01 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:40 +0000 UTC }] - Dec 14 15:57:11.875: INFO: ss-2 kiedai4iakae-2 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:40 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:40 +0000 UTC }] - Dec 14 15:57:11.875: INFO: - Dec 14 15:57:11.875: INFO: StatefulSet ss has not reached scale 0, at 3 - Dec 14 15:57:12.887: INFO: POD NODE PHASE GRACE CONDITIONS - Dec 14 15:57:12.887: INFO: ss-0 kiedai4iakae-3 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:19 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:19 +0000 UTC }] - Dec 14 15:57:12.888: INFO: ss-2 kiedai4iakae-2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:40 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:57:02 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 15:56:40 +0000 UTC }] - Dec 14 15:57:12.888: INFO: - Dec 14 15:57:12.888: INFO: StatefulSet ss has not reached scale 0, at 2 - Dec 14 15:57:13.894: INFO: Verifying statefulset ss doesn't scale past 0 for another 7.979160546s - Dec 14 15:57:14.904: INFO: Verifying statefulset ss doesn't scale past 0 for another 6.972597377s - Dec 14 15:57:15.915: INFO: Verifying statefulset ss doesn't scale past 0 for another 5.962082548s - Dec 14 15:57:16.929: INFO: Verifying statefulset ss doesn't scale past 0 for another 4.951435285s - Dec 14 15:57:17.935: INFO: Verifying statefulset ss doesn't scale past 0 for another 3.938553521s - Dec 14 15:57:18.942: INFO: Verifying statefulset ss doesn't scale past 0 for another 2.931382197s - Dec 14 15:57:19.948: INFO: Verifying statefulset ss doesn't scale past 0 for another 1.92418479s - Dec 14 15:57:20.958: INFO: Verifying statefulset ss doesn't scale past 0 for another 919.133634ms - STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-6999 12/14/22 15:57:21.959 - Dec 14 15:57:21.967: INFO: Scaling statefulset ss to 0 - Dec 14 15:57:22.000: INFO: Waiting for statefulset status.replicas updated to 0 - [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:122 - Dec 14 15:57:22.006: INFO: Deleting all statefulset in ns statefulset-6999 - Dec 14 15:57:22.012: INFO: Scaling statefulset ss to 0 - Dec 14 15:57:22.040: INFO: Waiting for statefulset status.replicas updated to 0 - Dec 14 15:57:22.044: INFO: Deleting statefulset ss - [AfterEach] [sig-apps] StatefulSet + STEP: Creating a kubernetes client 02/25/23 09:59:15.104 + Feb 25 09:59:15.104: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-lifecycle-hook 02/25/23 09:59:15.106 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:59:15.134 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:59:15.138 + [BeforeEach] when create a pod with lifecycle hook + test/e2e/common/node/lifecycle_hook.go:55 + STEP: create the container to handle the HTTPGet hook request. 02/25/23 09:59:15.154 + Feb 25 09:59:15.169: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-2342" to be "running and ready" + Feb 25 09:59:15.185: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 15.236413ms + Feb 25 09:59:15.185: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) + Feb 25 09:59:17.195: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02569601s + Feb 25 09:59:17.195: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) + Feb 25 09:59:19.192: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 4.022451563s + Feb 25 09:59:19.192: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) + Feb 25 09:59:19.192: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" + [It] should execute poststart exec hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:97 + STEP: create the pod with lifecycle hook 02/25/23 09:59:19.198 + Feb 25 09:59:19.209: INFO: Waiting up to 5m0s for pod "pod-with-poststart-exec-hook" in namespace "container-lifecycle-hook-2342" to be "running and ready" + Feb 25 09:59:19.216: INFO: Pod "pod-with-poststart-exec-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 6.597926ms + Feb 25 09:59:19.216: INFO: The phase of Pod pod-with-poststart-exec-hook is Pending, waiting for it to be Running (with Ready = true) + Feb 25 09:59:21.226: INFO: Pod "pod-with-poststart-exec-hook": Phase="Running", Reason="", readiness=true. Elapsed: 2.016274317s + Feb 25 09:59:21.226: INFO: The phase of Pod pod-with-poststart-exec-hook is Running (Ready = true) + Feb 25 09:59:21.226: INFO: Pod "pod-with-poststart-exec-hook" satisfied condition "running and ready" + STEP: check poststart hook 02/25/23 09:59:21.232 + STEP: delete the pod with lifecycle hook 02/25/23 09:59:21.244 + Feb 25 09:59:21.257: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear + Feb 25 09:59:21.263: INFO: Pod pod-with-poststart-exec-hook still exists + Feb 25 09:59:23.264: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear + Feb 25 09:59:23.272: INFO: Pod pod-with-poststart-exec-hook still exists + Feb 25 09:59:25.264: INFO: Waiting for pod pod-with-poststart-exec-hook to disappear + Feb 25 09:59:25.270: INFO: Pod pod-with-poststart-exec-hook no longer exists + [AfterEach] [sig-node] Container Lifecycle Hook test/e2e/framework/framework.go:187 - Dec 14 15:57:22.065: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "statefulset-6999" for this suite. 12/14/22 15:57:22.078 + Feb 25 09:59:25.270: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-lifecycle-hook-2342" for this suite. 02/25/23 09:59:25.278 << End Captured GinkgoWriter Output ------------------------------ -SSSSS +SSSSSSS ------------------------------ -[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] - should be able to convert from CR v1 to CR v2 [Conformance] - test/e2e/apimachinery/crd_conversion_webhook.go:149 -[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] +[sig-api-machinery] ResourceQuota + should verify ResourceQuota with terminating scopes. [Conformance] + test/e2e/apimachinery/resource_quota.go:680 +[BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:57:22.097 -Dec 14 15:57:22.097: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename crd-webhook 12/14/22 15:57:22.103 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:57:22.133 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:57:22.137 -[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/crd_conversion_webhook.go:128 -STEP: Setting up server cert 12/14/22 15:57:22.14 -STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication 12/14/22 15:57:24.076 -STEP: Deploying the custom resource conversion webhook pod 12/14/22 15:57:24.096 -STEP: Wait for the deployment to be ready 12/14/22 15:57:24.126 -Dec 14 15:57:24.167: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set -STEP: Deploying the webhook service 12/14/22 15:57:26.202 -STEP: Verifying the service has paired with the endpoint 12/14/22 15:57:26.227 -Dec 14 15:57:27.228: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1 -[It] should be able to convert from CR v1 to CR v2 [Conformance] - test/e2e/apimachinery/crd_conversion_webhook.go:149 -Dec 14 15:57:27.233: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Creating a v1 custom resource 12/14/22 15:57:29.988 -STEP: v2 custom resource should be converted 12/14/22 15:57:29.998 -[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 09:59:25.292 +Feb 25 09:59:25.293: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename resourcequota 02/25/23 09:59:25.295 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:59:25.332 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:59:25.336 +[It] should verify ResourceQuota with terminating scopes. [Conformance] + test/e2e/apimachinery/resource_quota.go:680 +STEP: Creating a ResourceQuota with terminating scope 02/25/23 09:59:25.342 +STEP: Ensuring ResourceQuota status is calculated 02/25/23 09:59:25.353 +STEP: Creating a ResourceQuota with not terminating scope 02/25/23 09:59:27.363 +STEP: Ensuring ResourceQuota status is calculated 02/25/23 09:59:27.373 +STEP: Creating a long running pod 02/25/23 09:59:29.379 +STEP: Ensuring resource quota with not terminating scope captures the pod usage 02/25/23 09:59:29.407 +STEP: Ensuring resource quota with terminating scope ignored the pod usage 02/25/23 09:59:31.424 +STEP: Deleting the pod 02/25/23 09:59:33.431 +STEP: Ensuring resource quota status released the pod usage 02/25/23 09:59:33.458 +STEP: Creating a terminating pod 02/25/23 09:59:35.471 +STEP: Ensuring resource quota with terminating scope captures the pod usage 02/25/23 09:59:35.494 +STEP: Ensuring resource quota with not terminating scope ignored the pod usage 02/25/23 09:59:37.507 +STEP: Deleting the pod 02/25/23 09:59:39.515 +STEP: Ensuring resource quota status released the pod usage 02/25/23 09:59:39.539 +[AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 -Dec 14 15:57:30.535: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-webhook-9792" for this suite. 12/14/22 15:57:30.551 -[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/crd_conversion_webhook.go:139 -{"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert from CR v1 to CR v2 [Conformance]","completed":45,"skipped":775,"failed":0} +Feb 25 09:59:41.549: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-1857" for this suite. 02/25/23 09:59:41.557 +{"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes. [Conformance]","completed":46,"skipped":1075,"failed":0} ------------------------------ -• [SLOW TEST] [8.528 seconds] -[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] +• [SLOW TEST] [16.276 seconds] +[sig-api-machinery] ResourceQuota test/e2e/apimachinery/framework.go:23 - should be able to convert from CR v1 to CR v2 [Conformance] - test/e2e/apimachinery/crd_conversion_webhook.go:149 + should verify ResourceQuota with terminating scopes. [Conformance] + test/e2e/apimachinery/resource_quota.go:680 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] + [BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:57:22.097 - Dec 14 15:57:22.097: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename crd-webhook 12/14/22 15:57:22.103 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:57:22.133 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:57:22.137 - [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/crd_conversion_webhook.go:128 - STEP: Setting up server cert 12/14/22 15:57:22.14 - STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication 12/14/22 15:57:24.076 - STEP: Deploying the custom resource conversion webhook pod 12/14/22 15:57:24.096 - STEP: Wait for the deployment to be ready 12/14/22 15:57:24.126 - Dec 14 15:57:24.167: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set - STEP: Deploying the webhook service 12/14/22 15:57:26.202 - STEP: Verifying the service has paired with the endpoint 12/14/22 15:57:26.227 - Dec 14 15:57:27.228: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1 - [It] should be able to convert from CR v1 to CR v2 [Conformance] - test/e2e/apimachinery/crd_conversion_webhook.go:149 - Dec 14 15:57:27.233: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Creating a v1 custom resource 12/14/22 15:57:29.988 - STEP: v2 custom resource should be converted 12/14/22 15:57:29.998 - [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 09:59:25.292 + Feb 25 09:59:25.293: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename resourcequota 02/25/23 09:59:25.295 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:59:25.332 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:59:25.336 + [It] should verify ResourceQuota with terminating scopes. [Conformance] + test/e2e/apimachinery/resource_quota.go:680 + STEP: Creating a ResourceQuota with terminating scope 02/25/23 09:59:25.342 + STEP: Ensuring ResourceQuota status is calculated 02/25/23 09:59:25.353 + STEP: Creating a ResourceQuota with not terminating scope 02/25/23 09:59:27.363 + STEP: Ensuring ResourceQuota status is calculated 02/25/23 09:59:27.373 + STEP: Creating a long running pod 02/25/23 09:59:29.379 + STEP: Ensuring resource quota with not terminating scope captures the pod usage 02/25/23 09:59:29.407 + STEP: Ensuring resource quota with terminating scope ignored the pod usage 02/25/23 09:59:31.424 + STEP: Deleting the pod 02/25/23 09:59:33.431 + STEP: Ensuring resource quota status released the pod usage 02/25/23 09:59:33.458 + STEP: Creating a terminating pod 02/25/23 09:59:35.471 + STEP: Ensuring resource quota with terminating scope captures the pod usage 02/25/23 09:59:35.494 + STEP: Ensuring resource quota with not terminating scope ignored the pod usage 02/25/23 09:59:37.507 + STEP: Deleting the pod 02/25/23 09:59:39.515 + STEP: Ensuring resource quota status released the pod usage 02/25/23 09:59:39.539 + [AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 - Dec 14 15:57:30.535: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "crd-webhook-9792" for this suite. 12/14/22 15:57:30.551 - [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/crd_conversion_webhook.go:139 + Feb 25 09:59:41.549: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "resourcequota-1857" for this suite. 02/25/23 09:59:41.557 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-cli] Kubectl client Guestbook application - should create and stop a working application [Conformance] - test/e2e/kubectl/kubectl.go:392 -[BeforeEach] [sig-cli] Kubectl client +[sig-apps] Deployment + deployment should support proportional scaling [Conformance] + test/e2e/apps/deployment.go:160 +[BeforeEach] [sig-apps] Deployment test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:57:30.644 -Dec 14 15:57:30.644: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 15:57:30.646 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:57:30.694 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:57:30.698 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 -[It] should create and stop a working application [Conformance] - test/e2e/kubectl/kubectl.go:392 -STEP: creating all guestbook components 12/14/22 15:57:30.706 -Dec 14 15:57:30.707: INFO: apiVersion: v1 -kind: Service -metadata: - name: agnhost-replica - labels: - app: agnhost - role: replica - tier: backend -spec: - ports: - - port: 6379 - selector: - app: agnhost - role: replica - tier: backend - -Dec 14 15:57:30.707: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 create -f -' -Dec 14 15:57:32.359: INFO: stderr: "" -Dec 14 15:57:32.359: INFO: stdout: "service/agnhost-replica created\n" -Dec 14 15:57:32.359: INFO: apiVersion: v1 -kind: Service -metadata: - name: agnhost-primary - labels: - app: agnhost - role: primary - tier: backend -spec: - ports: - - port: 6379 - targetPort: 6379 - selector: - app: agnhost - role: primary - tier: backend - -Dec 14 15:57:32.360: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 create -f -' -Dec 14 15:57:33.973: INFO: stderr: "" -Dec 14 15:57:33.973: INFO: stdout: "service/agnhost-primary created\n" -Dec 14 15:57:33.981: INFO: apiVersion: v1 -kind: Service -metadata: - name: frontend - labels: - app: guestbook - tier: frontend -spec: - # if your cluster supports it, uncomment the following to automatically create - # an external load-balanced IP for the frontend service. - # type: LoadBalancer - ports: - - port: 80 - selector: - app: guestbook - tier: frontend - -Dec 14 15:57:33.982: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 create -f -' -Dec 14 15:57:34.427: INFO: stderr: "" -Dec 14 15:57:34.428: INFO: stdout: "service/frontend created\n" -Dec 14 15:57:34.428: INFO: apiVersion: apps/v1 -kind: Deployment -metadata: - name: frontend -spec: - replicas: 3 - selector: - matchLabels: - app: guestbook - tier: frontend - template: - metadata: - labels: - app: guestbook - tier: frontend - spec: - containers: - - name: guestbook-frontend - image: registry.k8s.io/e2e-test-images/agnhost:2.40 - args: [ "guestbook", "--backend-port", "6379" ] - resources: - requests: - cpu: 100m - memory: 100Mi - ports: - - containerPort: 80 - -Dec 14 15:57:34.431: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 create -f -' -Dec 14 15:57:34.774: INFO: stderr: "" -Dec 14 15:57:34.774: INFO: stdout: "deployment.apps/frontend created\n" -Dec 14 15:57:34.777: INFO: apiVersion: apps/v1 -kind: Deployment -metadata: - name: agnhost-primary -spec: - replicas: 1 - selector: - matchLabels: - app: agnhost - role: primary - tier: backend - template: - metadata: - labels: - app: agnhost - role: primary - tier: backend - spec: - containers: - - name: primary - image: registry.k8s.io/e2e-test-images/agnhost:2.40 - args: [ "guestbook", "--http-port", "6379" ] - resources: - requests: - cpu: 100m - memory: 100Mi - ports: - - containerPort: 6379 +STEP: Creating a kubernetes client 02/25/23 09:59:41.57 +Feb 25 09:59:41.570: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename deployment 02/25/23 09:59:41.573 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:59:41.6 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:59:41.605 +[BeforeEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:91 +[It] deployment should support proportional scaling [Conformance] + test/e2e/apps/deployment.go:160 +Feb 25 09:59:41.611: INFO: Creating deployment "webserver-deployment" +Feb 25 09:59:41.621: INFO: Waiting for observed generation 1 +Feb 25 09:59:43.645: INFO: Waiting for all required pods to come up +Feb 25 09:59:43.659: INFO: Pod name httpd: Found 10 pods out of 10 +STEP: ensuring each pod is running 02/25/23 09:59:43.659 +Feb 25 09:59:43.660: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-d5cw8" in namespace "deployment-9570" to be "running" +Feb 25 09:59:43.660: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-bd6lz" in namespace "deployment-9570" to be "running" +Feb 25 09:59:43.661: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-c9jcc" in namespace "deployment-9570" to be "running" +Feb 25 09:59:43.661: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-ck8xn" in namespace "deployment-9570" to be "running" +Feb 25 09:59:43.662: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-ckmhn" in namespace "deployment-9570" to be "running" +Feb 25 09:59:43.662: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-qhh6v" in namespace "deployment-9570" to be "running" +Feb 25 09:59:43.662: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-d9w9f" in namespace "deployment-9570" to be "running" +Feb 25 09:59:43.662: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-gn54c" in namespace "deployment-9570" to be "running" +Feb 25 09:59:43.662: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-jnkp2" in namespace "deployment-9570" to be "running" +Feb 25 09:59:43.662: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-lqzcl" in namespace "deployment-9570" to be "running" +Feb 25 09:59:43.668: INFO: Pod "webserver-deployment-845c8977d9-c9jcc": Phase="Pending", Reason="", readiness=false. Elapsed: 7.232414ms +Feb 25 09:59:43.671: INFO: Pod "webserver-deployment-845c8977d9-d5cw8": Phase="Pending", Reason="", readiness=false. Elapsed: 10.869364ms +Feb 25 09:59:43.672: INFO: Pod "webserver-deployment-845c8977d9-d9w9f": Phase="Pending", Reason="", readiness=false. Elapsed: 9.734337ms +Feb 25 09:59:43.674: INFO: Pod "webserver-deployment-845c8977d9-jnkp2": Phase="Pending", Reason="", readiness=false. Elapsed: 11.311142ms +Feb 25 09:59:43.675: INFO: Pod "webserver-deployment-845c8977d9-gn54c": Phase="Pending", Reason="", readiness=false. Elapsed: 11.698477ms +Feb 25 09:59:43.675: INFO: Pod "webserver-deployment-845c8977d9-lqzcl": Phase="Pending", Reason="", readiness=false. Elapsed: 12.004161ms +Feb 25 09:59:43.675: INFO: Pod "webserver-deployment-845c8977d9-ck8xn": Phase="Pending", Reason="", readiness=false. Elapsed: 14.017514ms +Feb 25 09:59:43.682: INFO: Pod "webserver-deployment-845c8977d9-bd6lz": Phase="Pending", Reason="", readiness=false. Elapsed: 21.326885ms +Feb 25 09:59:43.684: INFO: Pod "webserver-deployment-845c8977d9-qhh6v": Phase="Pending", Reason="", readiness=false. Elapsed: 22.017377ms +Feb 25 09:59:43.684: INFO: Pod "webserver-deployment-845c8977d9-ckmhn": Phase="Pending", Reason="", readiness=false. Elapsed: 22.754427ms +Feb 25 09:59:45.682: INFO: Pod "webserver-deployment-845c8977d9-c9jcc": Phase="Running", Reason="", readiness=true. Elapsed: 2.021071117s +Feb 25 09:59:45.683: INFO: Pod "webserver-deployment-845c8977d9-c9jcc" satisfied condition "running" +Feb 25 09:59:45.684: INFO: Pod "webserver-deployment-845c8977d9-d5cw8": Phase="Running", Reason="", readiness=true. Elapsed: 2.024009057s +Feb 25 09:59:45.684: INFO: Pod "webserver-deployment-845c8977d9-d5cw8" satisfied condition "running" +Feb 25 09:59:45.684: INFO: Pod "webserver-deployment-845c8977d9-gn54c": Phase="Running", Reason="", readiness=true. Elapsed: 2.021502662s +Feb 25 09:59:45.686: INFO: Pod "webserver-deployment-845c8977d9-gn54c" satisfied condition "running" +Feb 25 09:59:45.685: INFO: Pod "webserver-deployment-845c8977d9-d9w9f": Phase="Running", Reason="", readiness=true. Elapsed: 2.022247848s +Feb 25 09:59:45.686: INFO: Pod "webserver-deployment-845c8977d9-d9w9f" satisfied condition "running" +Feb 25 09:59:45.685: INFO: Pod "webserver-deployment-845c8977d9-jnkp2": Phase="Running", Reason="", readiness=true. Elapsed: 2.022492951s +Feb 25 09:59:45.686: INFO: Pod "webserver-deployment-845c8977d9-jnkp2" satisfied condition "running" +Feb 25 09:59:45.690: INFO: Pod "webserver-deployment-845c8977d9-lqzcl": Phase="Running", Reason="", readiness=true. Elapsed: 2.027084195s +Feb 25 09:59:45.690: INFO: Pod "webserver-deployment-845c8977d9-lqzcl" satisfied condition "running" +Feb 25 09:59:45.692: INFO: Pod "webserver-deployment-845c8977d9-ck8xn": Phase="Running", Reason="", readiness=true. Elapsed: 2.030651746s +Feb 25 09:59:45.692: INFO: Pod "webserver-deployment-845c8977d9-ck8xn" satisfied condition "running" +Feb 25 09:59:45.693: INFO: Pod "webserver-deployment-845c8977d9-bd6lz": Phase="Running", Reason="", readiness=true. Elapsed: 2.032439326s +Feb 25 09:59:45.693: INFO: Pod "webserver-deployment-845c8977d9-bd6lz" satisfied condition "running" +Feb 25 09:59:45.694: INFO: Pod "webserver-deployment-845c8977d9-qhh6v": Phase="Running", Reason="", readiness=true. Elapsed: 2.032050317s +Feb 25 09:59:45.694: INFO: Pod "webserver-deployment-845c8977d9-qhh6v" satisfied condition "running" +Feb 25 09:59:45.702: INFO: Pod "webserver-deployment-845c8977d9-ckmhn": Phase="Running", Reason="", readiness=true. Elapsed: 2.040641237s +Feb 25 09:59:45.702: INFO: Pod "webserver-deployment-845c8977d9-ckmhn" satisfied condition "running" +Feb 25 09:59:45.702: INFO: Waiting for deployment "webserver-deployment" to complete +Feb 25 09:59:45.715: INFO: Updating deployment "webserver-deployment" with a non-existent image +Feb 25 09:59:45.732: INFO: Updating deployment webserver-deployment +Feb 25 09:59:45.732: INFO: Waiting for observed generation 2 +Feb 25 09:59:47.747: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8 +Feb 25 09:59:47.754: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8 +Feb 25 09:59:47.761: INFO: Waiting for the first rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas +Feb 25 09:59:47.778: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0 +Feb 25 09:59:47.778: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5 +Feb 25 09:59:47.783: INFO: Waiting for the second rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas +Feb 25 09:59:47.795: INFO: Verifying that deployment "webserver-deployment" has minimum required number of available replicas +Feb 25 09:59:47.795: INFO: Scaling up the deployment "webserver-deployment" from 10 to 30 +Feb 25 09:59:47.814: INFO: Updating deployment webserver-deployment +Feb 25 09:59:47.814: INFO: Waiting for the replicasets of deployment "webserver-deployment" to have desired number of replicas +Feb 25 09:59:47.834: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20 +Feb 25 09:59:50.063: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13 +[AfterEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:84 +Feb 25 09:59:50.150: INFO: Deployment "webserver-deployment": +&Deployment{ObjectMeta:{webserver-deployment deployment-9570 c0595620-6cd4-454f-be37-d525686a1712 12405 3 2023-02-25 09:59:41 +0000 UTC map[name:httpd] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2023-02-25 09:59:47 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:unavailableReplicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*30,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc002a634c8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:33,UpdatedReplicas:13,AvailableReplicas:8,UnavailableReplicas:25,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2023-02-25 09:59:47 +0000 UTC,LastTransitionTime:2023-02-25 09:59:47 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "webserver-deployment-69b7448995" is progressing.,LastUpdateTime:2023-02-25 09:59:48 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,},},ReadyReplicas:8,CollisionCount:nil,},} + +Feb 25 09:59:50.242: INFO: New ReplicaSet "webserver-deployment-69b7448995" of Deployment "webserver-deployment": +&ReplicaSet{ObjectMeta:{webserver-deployment-69b7448995 deployment-9570 c5eaf26a-2a6e-441f-b269-1dfd6870022c 12396 3 2023-02-25 09:59:45 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment webserver-deployment c0595620-6cd4-454f-be37-d525686a1712 0xc002a638d7 0xc002a638d8}] [] [{kube-controller-manager Update apps/v1 2023-02-25 09:59:47 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c0595620-6cd4-454f-be37-d525686a1712\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*13,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 69b7448995,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc002a63978 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:13,FullyLabeledReplicas:13,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} +Feb 25 09:59:50.242: INFO: All old ReplicaSets of Deployment "webserver-deployment": +Feb 25 09:59:50.242: INFO: &ReplicaSet{ObjectMeta:{webserver-deployment-845c8977d9 deployment-9570 629b2891-0215-442f-820d-aa24812ead01 12398 3 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment webserver-deployment c0595620-6cd4-454f-be37-d525686a1712 0xc002a639d7 0xc002a639d8}] [] [{kube-controller-manager Update apps/v1 2023-02-25 09:59:47 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c0595620-6cd4-454f-be37-d525686a1712\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*20,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 845c8977d9,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc002a63a68 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:20,FullyLabeledReplicas:20,ObservedGeneration:3,ReadyReplicas:8,AvailableReplicas:8,Conditions:[]ReplicaSetCondition{},},} +Feb 25 09:59:50.308: INFO: Pod "webserver-deployment-69b7448995-2mtvt" is not available: +&Pod{ObjectMeta:{webserver-deployment-69b7448995-2mtvt webserver-deployment-69b7448995- deployment-9570 1c136676-5d8a-4730-bd6f-6480ca2e8ace 12383 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc0032487d7 0xc0032487d8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-r78gm,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r78gm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.312: INFO: Pod "webserver-deployment-69b7448995-5zzd2" is not available: +&Pod{ObjectMeta:{webserver-deployment-69b7448995-5zzd2 webserver-deployment-69b7448995- deployment-9570 f7339f07-90d8-465b-87ae-46e63b58f9bb 12397 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc0032489c7 0xc0032489c8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-9tzc7,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9tzc7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.313: INFO: Pod "webserver-deployment-69b7448995-cckj5" is not available: +&Pod{ObjectMeta:{webserver-deployment-69b7448995-cckj5 webserver-deployment-69b7448995- deployment-9570 7e8f26db-39d4-4a58-9e1b-85998c57fe10 12388 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc003248b40 0xc003248b41}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-fnscx,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fnscx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.314: INFO: Pod "webserver-deployment-69b7448995-fb7wz" is not available: +&Pod{ObjectMeta:{webserver-deployment-69b7448995-fb7wz webserver-deployment-69b7448995- deployment-9570 f1cec890-67df-4a2b-bbc4-77a1d7c0313e 12293 0 2023-02-25 09:59:45 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc003248d27 0xc003248d28}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:45 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:46 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-98cb4,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-98cb4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:46 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.323: INFO: Pod "webserver-deployment-69b7448995-g4gbh" is not available: +&Pod{ObjectMeta:{webserver-deployment-69b7448995-g4gbh webserver-deployment-69b7448995- deployment-9570 6d287f0d-243d-4ec7-9269-330fd8a3546b 12354 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc003248f17 0xc003248f18}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-s9xfw,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s9xfw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.324: INFO: Pod "webserver-deployment-69b7448995-hpt7h" is not available: +&Pod{ObjectMeta:{webserver-deployment-69b7448995-hpt7h webserver-deployment-69b7448995- deployment-9570 c1b19c50-58df-471b-a1d2-bad7ee222c23 12276 0 2023-02-25 09:59:45 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc003249107 0xc003249108}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:45 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:45 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-56cjl,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-56cjl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.325: INFO: Pod "webserver-deployment-69b7448995-kj95b" is not available: +&Pod{ObjectMeta:{webserver-deployment-69b7448995-kj95b webserver-deployment-69b7448995- deployment-9570 2c4806dc-4d43-4f31-be3b-7f674dafc383 12360 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc0032492f7 0xc0032492f8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-6xgn6,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6xgn6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.330: INFO: Pod "webserver-deployment-69b7448995-lb5bv" is not available: +&Pod{ObjectMeta:{webserver-deployment-69b7448995-lb5bv webserver-deployment-69b7448995- deployment-9570 05ecc179-4060-49c4-a1f4-539b9ec6cace 12384 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc0032494e7 0xc0032494e8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-j95rb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j95rb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.330: INFO: Pod "webserver-deployment-69b7448995-r5xr9" is not available: +&Pod{ObjectMeta:{webserver-deployment-69b7448995-r5xr9 webserver-deployment-69b7448995- deployment-9570 efda9be2-4766-4f5a-a79e-61179058af8b 12434 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc0032496d7 0xc0032496d8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:49 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-fgbt7,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fgbt7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.333: INFO: Pod "webserver-deployment-69b7448995-smzgk" is not available: +&Pod{ObjectMeta:{webserver-deployment-69b7448995-smzgk webserver-deployment-69b7448995- deployment-9570 ffef5433-c3b0-4dd1-a483-5026937e982d 12301 0 2023-02-25 09:59:46 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc0032498c7 0xc0032498c8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:46 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:46 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-l4jlf,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l4jlf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:46 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:46 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:46 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:46 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:,StartTime:2023-02-25 09:59:46 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.333: INFO: Pod "webserver-deployment-69b7448995-twl7x" is not available: +&Pod{ObjectMeta:{webserver-deployment-69b7448995-twl7x webserver-deployment-69b7448995- deployment-9570 c76161c2-f1db-453d-906e-584aac8384eb 12392 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc003249ab7 0xc003249ab8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-lwt4b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lwt4b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.339: INFO: Pod "webserver-deployment-69b7448995-w8bt7" is not available: +&Pod{ObjectMeta:{webserver-deployment-69b7448995-w8bt7 webserver-deployment-69b7448995- deployment-9570 cb43181e-10f2-4c54-b26d-86335d288017 12281 0 2023-02-25 09:59:45 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc003249ca7 0xc003249ca8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:45 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:45 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-sdd2k,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sdd2k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:,StartTime:2023-02-25 09:59:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.340: INFO: Pod "webserver-deployment-69b7448995-xn2vh" is not available: +&Pod{ObjectMeta:{webserver-deployment-69b7448995-xn2vh webserver-deployment-69b7448995- deployment-9570 b991992b-41f1-41e7-82a7-3711983cda06 12275 0 2023-02-25 09:59:45 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc003249e97 0xc003249e98}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:45 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:45 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-d96zw,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d96zw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:,StartTime:2023-02-25 09:59:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.340: INFO: Pod "webserver-deployment-845c8977d9-2x6m6" is not available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-2x6m6 webserver-deployment-845c8977d9- deployment-9570 c015d3ff-14f5-48f3-a2e6-56a320e83364 12390 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003058087 0xc003058088}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-qn6gg,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qn6gg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.344: INFO: Pod "webserver-deployment-845c8977d9-6xcgk" is not available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-6xcgk webserver-deployment-845c8977d9- deployment-9570 4c62de82-5650-4f82-bac3-94bee3e82f03 12393 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc0030581f0 0xc0030581f1}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-hsh58,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hsh58,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.351: INFO: Pod "webserver-deployment-845c8977d9-77gk6" is not available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-77gk6 webserver-deployment-845c8977d9- deployment-9570 874fde19-74ec-4e1f-a483-29cd4d9ffa64 12403 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc0030583b7 0xc0030583b8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-mcxqm,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mcxqm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.352: INFO: Pod "webserver-deployment-845c8977d9-96wtz" is not available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-96wtz webserver-deployment-845c8977d9- deployment-9570 59bb96ec-53a4-4719-9f1b-204b7fbe6c36 12349 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003058587 0xc003058588}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-cwd2k,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cwd2k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.353: INFO: Pod "webserver-deployment-845c8977d9-9chbw" is not available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-9chbw webserver-deployment-845c8977d9- deployment-9570 92da79a1-6a77-400d-b52f-739dfb40a7e6 12335 0 2023-02-25 09:59:47 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003058757 0xc003058758}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:47 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:47 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-dvv52,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dvv52,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:47 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:47 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:47 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:47 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.353: INFO: Pod "webserver-deployment-845c8977d9-bd6lz" is available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-bd6lz webserver-deployment-845c8977d9- deployment-9570 6bee04e4-f18a-4ecb-928c-eab449f58ff5 12229 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003058927 0xc003058928}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.224\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-qzmmg,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qzmmg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:10.233.64.224,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://4abec4088dd7f45a8d07741449bbdcda323b1b104a6b3e3aac612214ab9f254c,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.224,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.354: INFO: Pod "webserver-deployment-845c8977d9-c9jcc" is available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-c9jcc webserver-deployment-845c8977d9- deployment-9570 51d46c16-41d4-4f5e-aa4e-e85e0c2a6c62 12235 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003058b17 0xc003058b18}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.83\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-wxzkx,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wxzkx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:10.233.65.83,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://1f1ea883ede1a3b7d04ef6d6dc15f9d91f5a23dbb37776eee35e4a760043eb87,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.83,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.354: INFO: Pod "webserver-deployment-845c8977d9-ck8xn" is available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-ck8xn webserver-deployment-845c8977d9- deployment-9570 fd954b1e-d4a1-492f-adcd-64d84826337b 12238 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003058d07 0xc003058d08}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.205\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-46zhh,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-46zhh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:10.233.65.205,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://bc982f63b06bb0a1a8bf7f3510ed2ce500c10b9875930d33142ca046f1e71570,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.205,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.354: INFO: Pod "webserver-deployment-845c8977d9-ckmhn" is available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-ckmhn webserver-deployment-845c8977d9- deployment-9570 981613e0-a706-4035-96d1-c7b6d3f29506 12227 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003058ef7 0xc003058ef8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.95\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-brb9t,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-brb9t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:10.233.64.95,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://3b602164883b7305eb49e1a195847c927a0300e6347a7798018c6159099ae3d5,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.95,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.355: INFO: Pod "webserver-deployment-845c8977d9-d5cw8" is available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-d5cw8 webserver-deployment-845c8977d9- deployment-9570 ce667004-92a2-4a94-a9cf-1e5b17be0de0 12226 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc0030590e7 0xc0030590e8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.96\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-pxf44,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pxf44,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:10.233.64.96,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://0a98f0f392f519251ccaf685893f1be5984853bad78de06700f10d1bebbc271d,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.96,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.360: INFO: Pod "webserver-deployment-845c8977d9-gn54c" is available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-gn54c webserver-deployment-845c8977d9- deployment-9570 aa3df65d-eca4-491d-8dcf-0bf6c0ddc6e7 12230 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc0030592d7 0xc0030592d8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.61\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-xztft,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xztft,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:10.233.65.61,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://9fcd0df0077547e16d67ec3c975044bd298aafcac9c1c94948c3dacba9048325,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.61,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.360: INFO: Pod "webserver-deployment-845c8977d9-jnkp2" is available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-jnkp2 webserver-deployment-845c8977d9- deployment-9570 56ed147d-2f8e-423d-914c-a80ce1c8f249 12219 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc0030594c7 0xc0030594c8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.142\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-5w675,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5w675,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:10.233.66.142,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://0af54bf851c09e577d02ea22cbe53a016306dfb7681b65a563bc80325acd505d,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.142,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.360: INFO: Pod "webserver-deployment-845c8977d9-kww24" is not available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-kww24 webserver-deployment-845c8977d9- deployment-9570 39d5cff6-4082-4d68-b609-d7e39e3a90b4 12457 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc0030596b7 0xc0030596b8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:50 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-vqbtf,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vqbtf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.361: INFO: Pod "webserver-deployment-845c8977d9-lqzcl" is available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-lqzcl webserver-deployment-845c8977d9- deployment-9570 4678a213-a27b-47b1-b043-d0f9cd25c3d8 12222 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003059887 0xc003059888}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.61\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-vnkmv,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vnkmv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:10.233.66.61,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://48cde622b9e168bd90d228dc98cc4bc490c4dd6bae5163712e77077fbdde531f,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.61,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.361: INFO: Pod "webserver-deployment-845c8977d9-m6q4v" is not available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-m6q4v webserver-deployment-845c8977d9- deployment-9570 049cb947-a0ad-47f7-9c1f-52efcc5fd328 12391 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003059be7 0xc003059be8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-796w5,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-796w5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.361: INFO: Pod "webserver-deployment-845c8977d9-mvmv2" is not available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-mvmv2 webserver-deployment-845c8977d9- deployment-9570 ef03a776-e514-4af1-be4d-01e53f4b6c75 12419 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003059d50 0xc003059d51}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-kkx7b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kkx7b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.362: INFO: Pod "webserver-deployment-845c8977d9-prvk8" is not available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-prvk8 webserver-deployment-845c8977d9- deployment-9570 ebfb6305-1196-4bd8-8e2f-314935e3b098 12420 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003059f17 0xc003059f18}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-4pwz4,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4pwz4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.363: INFO: Pod "webserver-deployment-845c8977d9-tsjd2" is not available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-tsjd2 webserver-deployment-845c8977d9- deployment-9570 776e9a27-7e3d-4631-ab1a-774d98c49743 12342 0 2023-02-25 09:59:47 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc00285e377 0xc00285e378}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:47 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-r55kc,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r55kc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.363: INFO: Pod "webserver-deployment-845c8977d9-vbf2v" is not available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-vbf2v webserver-deployment-845c8977d9- deployment-9570 b81c118c-0c53-4002-8500-983df02a5200 12444 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc00285e667 0xc00285e668}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:49 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-pq2cf,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pq2cf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 09:59:50.364: INFO: Pod "webserver-deployment-845c8977d9-xfdc9" is not available: +&Pod{ObjectMeta:{webserver-deployment-845c8977d9-xfdc9 webserver-deployment-845c8977d9- deployment-9570 aae3f967-2d08-404c-b3cd-eac3361c80af 12400 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc00285ec67 0xc00285ec68}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-gbmkp,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gbmkp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +[AfterEach] [sig-apps] Deployment + test/e2e/framework/framework.go:187 +Feb 25 09:59:50.364: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "deployment-9570" for this suite. 02/25/23 09:59:50.463 +{"msg":"PASSED [sig-apps] Deployment deployment should support proportional scaling [Conformance]","completed":47,"skipped":1075,"failed":0} +------------------------------ +• [SLOW TEST] [8.932 seconds] +[sig-apps] Deployment +test/e2e/apps/framework.go:23 + deployment should support proportional scaling [Conformance] + test/e2e/apps/deployment.go:160 -Dec 14 15:57:34.778: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 create -f -' -Dec 14 15:57:35.134: INFO: stderr: "" -Dec 14 15:57:35.134: INFO: stdout: "deployment.apps/agnhost-primary created\n" -Dec 14 15:57:35.134: INFO: apiVersion: apps/v1 -kind: Deployment -metadata: - name: agnhost-replica -spec: - replicas: 2 - selector: - matchLabels: - app: agnhost - role: replica - tier: backend - template: - metadata: - labels: - app: agnhost - role: replica - tier: backend - spec: - containers: - - name: replica - image: registry.k8s.io/e2e-test-images/agnhost:2.40 - args: [ "guestbook", "--replicaof", "agnhost-primary", "--http-port", "6379" ] - resources: - requests: - cpu: 100m - memory: 100Mi - ports: - - containerPort: 6379 + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-apps] Deployment + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 09:59:41.57 + Feb 25 09:59:41.570: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename deployment 02/25/23 09:59:41.573 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:59:41.6 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:59:41.605 + [BeforeEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:91 + [It] deployment should support proportional scaling [Conformance] + test/e2e/apps/deployment.go:160 + Feb 25 09:59:41.611: INFO: Creating deployment "webserver-deployment" + Feb 25 09:59:41.621: INFO: Waiting for observed generation 1 + Feb 25 09:59:43.645: INFO: Waiting for all required pods to come up + Feb 25 09:59:43.659: INFO: Pod name httpd: Found 10 pods out of 10 + STEP: ensuring each pod is running 02/25/23 09:59:43.659 + Feb 25 09:59:43.660: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-d5cw8" in namespace "deployment-9570" to be "running" + Feb 25 09:59:43.660: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-bd6lz" in namespace "deployment-9570" to be "running" + Feb 25 09:59:43.661: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-c9jcc" in namespace "deployment-9570" to be "running" + Feb 25 09:59:43.661: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-ck8xn" in namespace "deployment-9570" to be "running" + Feb 25 09:59:43.662: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-ckmhn" in namespace "deployment-9570" to be "running" + Feb 25 09:59:43.662: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-qhh6v" in namespace "deployment-9570" to be "running" + Feb 25 09:59:43.662: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-d9w9f" in namespace "deployment-9570" to be "running" + Feb 25 09:59:43.662: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-gn54c" in namespace "deployment-9570" to be "running" + Feb 25 09:59:43.662: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-jnkp2" in namespace "deployment-9570" to be "running" + Feb 25 09:59:43.662: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-lqzcl" in namespace "deployment-9570" to be "running" + Feb 25 09:59:43.668: INFO: Pod "webserver-deployment-845c8977d9-c9jcc": Phase="Pending", Reason="", readiness=false. Elapsed: 7.232414ms + Feb 25 09:59:43.671: INFO: Pod "webserver-deployment-845c8977d9-d5cw8": Phase="Pending", Reason="", readiness=false. Elapsed: 10.869364ms + Feb 25 09:59:43.672: INFO: Pod "webserver-deployment-845c8977d9-d9w9f": Phase="Pending", Reason="", readiness=false. Elapsed: 9.734337ms + Feb 25 09:59:43.674: INFO: Pod "webserver-deployment-845c8977d9-jnkp2": Phase="Pending", Reason="", readiness=false. Elapsed: 11.311142ms + Feb 25 09:59:43.675: INFO: Pod "webserver-deployment-845c8977d9-gn54c": Phase="Pending", Reason="", readiness=false. Elapsed: 11.698477ms + Feb 25 09:59:43.675: INFO: Pod "webserver-deployment-845c8977d9-lqzcl": Phase="Pending", Reason="", readiness=false. Elapsed: 12.004161ms + Feb 25 09:59:43.675: INFO: Pod "webserver-deployment-845c8977d9-ck8xn": Phase="Pending", Reason="", readiness=false. Elapsed: 14.017514ms + Feb 25 09:59:43.682: INFO: Pod "webserver-deployment-845c8977d9-bd6lz": Phase="Pending", Reason="", readiness=false. Elapsed: 21.326885ms + Feb 25 09:59:43.684: INFO: Pod "webserver-deployment-845c8977d9-qhh6v": Phase="Pending", Reason="", readiness=false. Elapsed: 22.017377ms + Feb 25 09:59:43.684: INFO: Pod "webserver-deployment-845c8977d9-ckmhn": Phase="Pending", Reason="", readiness=false. Elapsed: 22.754427ms + Feb 25 09:59:45.682: INFO: Pod "webserver-deployment-845c8977d9-c9jcc": Phase="Running", Reason="", readiness=true. Elapsed: 2.021071117s + Feb 25 09:59:45.683: INFO: Pod "webserver-deployment-845c8977d9-c9jcc" satisfied condition "running" + Feb 25 09:59:45.684: INFO: Pod "webserver-deployment-845c8977d9-d5cw8": Phase="Running", Reason="", readiness=true. Elapsed: 2.024009057s + Feb 25 09:59:45.684: INFO: Pod "webserver-deployment-845c8977d9-d5cw8" satisfied condition "running" + Feb 25 09:59:45.684: INFO: Pod "webserver-deployment-845c8977d9-gn54c": Phase="Running", Reason="", readiness=true. Elapsed: 2.021502662s + Feb 25 09:59:45.686: INFO: Pod "webserver-deployment-845c8977d9-gn54c" satisfied condition "running" + Feb 25 09:59:45.685: INFO: Pod "webserver-deployment-845c8977d9-d9w9f": Phase="Running", Reason="", readiness=true. Elapsed: 2.022247848s + Feb 25 09:59:45.686: INFO: Pod "webserver-deployment-845c8977d9-d9w9f" satisfied condition "running" + Feb 25 09:59:45.685: INFO: Pod "webserver-deployment-845c8977d9-jnkp2": Phase="Running", Reason="", readiness=true. Elapsed: 2.022492951s + Feb 25 09:59:45.686: INFO: Pod "webserver-deployment-845c8977d9-jnkp2" satisfied condition "running" + Feb 25 09:59:45.690: INFO: Pod "webserver-deployment-845c8977d9-lqzcl": Phase="Running", Reason="", readiness=true. Elapsed: 2.027084195s + Feb 25 09:59:45.690: INFO: Pod "webserver-deployment-845c8977d9-lqzcl" satisfied condition "running" + Feb 25 09:59:45.692: INFO: Pod "webserver-deployment-845c8977d9-ck8xn": Phase="Running", Reason="", readiness=true. Elapsed: 2.030651746s + Feb 25 09:59:45.692: INFO: Pod "webserver-deployment-845c8977d9-ck8xn" satisfied condition "running" + Feb 25 09:59:45.693: INFO: Pod "webserver-deployment-845c8977d9-bd6lz": Phase="Running", Reason="", readiness=true. Elapsed: 2.032439326s + Feb 25 09:59:45.693: INFO: Pod "webserver-deployment-845c8977d9-bd6lz" satisfied condition "running" + Feb 25 09:59:45.694: INFO: Pod "webserver-deployment-845c8977d9-qhh6v": Phase="Running", Reason="", readiness=true. Elapsed: 2.032050317s + Feb 25 09:59:45.694: INFO: Pod "webserver-deployment-845c8977d9-qhh6v" satisfied condition "running" + Feb 25 09:59:45.702: INFO: Pod "webserver-deployment-845c8977d9-ckmhn": Phase="Running", Reason="", readiness=true. Elapsed: 2.040641237s + Feb 25 09:59:45.702: INFO: Pod "webserver-deployment-845c8977d9-ckmhn" satisfied condition "running" + Feb 25 09:59:45.702: INFO: Waiting for deployment "webserver-deployment" to complete + Feb 25 09:59:45.715: INFO: Updating deployment "webserver-deployment" with a non-existent image + Feb 25 09:59:45.732: INFO: Updating deployment webserver-deployment + Feb 25 09:59:45.732: INFO: Waiting for observed generation 2 + Feb 25 09:59:47.747: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8 + Feb 25 09:59:47.754: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8 + Feb 25 09:59:47.761: INFO: Waiting for the first rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas + Feb 25 09:59:47.778: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0 + Feb 25 09:59:47.778: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5 + Feb 25 09:59:47.783: INFO: Waiting for the second rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas + Feb 25 09:59:47.795: INFO: Verifying that deployment "webserver-deployment" has minimum required number of available replicas + Feb 25 09:59:47.795: INFO: Scaling up the deployment "webserver-deployment" from 10 to 30 + Feb 25 09:59:47.814: INFO: Updating deployment webserver-deployment + Feb 25 09:59:47.814: INFO: Waiting for the replicasets of deployment "webserver-deployment" to have desired number of replicas + Feb 25 09:59:47.834: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20 + Feb 25 09:59:50.063: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13 + [AfterEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:84 + Feb 25 09:59:50.150: INFO: Deployment "webserver-deployment": + &Deployment{ObjectMeta:{webserver-deployment deployment-9570 c0595620-6cd4-454f-be37-d525686a1712 12405 3 2023-02-25 09:59:41 +0000 UTC map[name:httpd] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2023-02-25 09:59:47 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:unavailableReplicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*30,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc002a634c8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:33,UpdatedReplicas:13,AvailableReplicas:8,UnavailableReplicas:25,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2023-02-25 09:59:47 +0000 UTC,LastTransitionTime:2023-02-25 09:59:47 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "webserver-deployment-69b7448995" is progressing.,LastUpdateTime:2023-02-25 09:59:48 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,},},ReadyReplicas:8,CollisionCount:nil,},} + + Feb 25 09:59:50.242: INFO: New ReplicaSet "webserver-deployment-69b7448995" of Deployment "webserver-deployment": + &ReplicaSet{ObjectMeta:{webserver-deployment-69b7448995 deployment-9570 c5eaf26a-2a6e-441f-b269-1dfd6870022c 12396 3 2023-02-25 09:59:45 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment webserver-deployment c0595620-6cd4-454f-be37-d525686a1712 0xc002a638d7 0xc002a638d8}] [] [{kube-controller-manager Update apps/v1 2023-02-25 09:59:47 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c0595620-6cd4-454f-be37-d525686a1712\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*13,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 69b7448995,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc002a63978 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:13,FullyLabeledReplicas:13,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} + Feb 25 09:59:50.242: INFO: All old ReplicaSets of Deployment "webserver-deployment": + Feb 25 09:59:50.242: INFO: &ReplicaSet{ObjectMeta:{webserver-deployment-845c8977d9 deployment-9570 629b2891-0215-442f-820d-aa24812ead01 12398 3 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment webserver-deployment c0595620-6cd4-454f-be37-d525686a1712 0xc002a639d7 0xc002a639d8}] [] [{kube-controller-manager Update apps/v1 2023-02-25 09:59:47 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c0595620-6cd4-454f-be37-d525686a1712\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*20,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 845c8977d9,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc002a63a68 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:20,FullyLabeledReplicas:20,ObservedGeneration:3,ReadyReplicas:8,AvailableReplicas:8,Conditions:[]ReplicaSetCondition{},},} + Feb 25 09:59:50.308: INFO: Pod "webserver-deployment-69b7448995-2mtvt" is not available: + &Pod{ObjectMeta:{webserver-deployment-69b7448995-2mtvt webserver-deployment-69b7448995- deployment-9570 1c136676-5d8a-4730-bd6f-6480ca2e8ace 12383 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc0032487d7 0xc0032487d8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-r78gm,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r78gm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.312: INFO: Pod "webserver-deployment-69b7448995-5zzd2" is not available: + &Pod{ObjectMeta:{webserver-deployment-69b7448995-5zzd2 webserver-deployment-69b7448995- deployment-9570 f7339f07-90d8-465b-87ae-46e63b58f9bb 12397 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc0032489c7 0xc0032489c8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-9tzc7,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9tzc7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.313: INFO: Pod "webserver-deployment-69b7448995-cckj5" is not available: + &Pod{ObjectMeta:{webserver-deployment-69b7448995-cckj5 webserver-deployment-69b7448995- deployment-9570 7e8f26db-39d4-4a58-9e1b-85998c57fe10 12388 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc003248b40 0xc003248b41}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-fnscx,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fnscx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.314: INFO: Pod "webserver-deployment-69b7448995-fb7wz" is not available: + &Pod{ObjectMeta:{webserver-deployment-69b7448995-fb7wz webserver-deployment-69b7448995- deployment-9570 f1cec890-67df-4a2b-bbc4-77a1d7c0313e 12293 0 2023-02-25 09:59:45 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc003248d27 0xc003248d28}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:45 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:46 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-98cb4,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-98cb4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:46 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.323: INFO: Pod "webserver-deployment-69b7448995-g4gbh" is not available: + &Pod{ObjectMeta:{webserver-deployment-69b7448995-g4gbh webserver-deployment-69b7448995- deployment-9570 6d287f0d-243d-4ec7-9269-330fd8a3546b 12354 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc003248f17 0xc003248f18}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-s9xfw,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-s9xfw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.324: INFO: Pod "webserver-deployment-69b7448995-hpt7h" is not available: + &Pod{ObjectMeta:{webserver-deployment-69b7448995-hpt7h webserver-deployment-69b7448995- deployment-9570 c1b19c50-58df-471b-a1d2-bad7ee222c23 12276 0 2023-02-25 09:59:45 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc003249107 0xc003249108}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:45 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:45 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-56cjl,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-56cjl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.325: INFO: Pod "webserver-deployment-69b7448995-kj95b" is not available: + &Pod{ObjectMeta:{webserver-deployment-69b7448995-kj95b webserver-deployment-69b7448995- deployment-9570 2c4806dc-4d43-4f31-be3b-7f674dafc383 12360 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc0032492f7 0xc0032492f8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-6xgn6,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6xgn6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.330: INFO: Pod "webserver-deployment-69b7448995-lb5bv" is not available: + &Pod{ObjectMeta:{webserver-deployment-69b7448995-lb5bv webserver-deployment-69b7448995- deployment-9570 05ecc179-4060-49c4-a1f4-539b9ec6cace 12384 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc0032494e7 0xc0032494e8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-j95rb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j95rb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.330: INFO: Pod "webserver-deployment-69b7448995-r5xr9" is not available: + &Pod{ObjectMeta:{webserver-deployment-69b7448995-r5xr9 webserver-deployment-69b7448995- deployment-9570 efda9be2-4766-4f5a-a79e-61179058af8b 12434 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc0032496d7 0xc0032496d8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:49 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-fgbt7,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fgbt7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.333: INFO: Pod "webserver-deployment-69b7448995-smzgk" is not available: + &Pod{ObjectMeta:{webserver-deployment-69b7448995-smzgk webserver-deployment-69b7448995- deployment-9570 ffef5433-c3b0-4dd1-a483-5026937e982d 12301 0 2023-02-25 09:59:46 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc0032498c7 0xc0032498c8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:46 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:46 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-l4jlf,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-l4jlf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:46 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:46 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:46 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:46 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:,StartTime:2023-02-25 09:59:46 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.333: INFO: Pod "webserver-deployment-69b7448995-twl7x" is not available: + &Pod{ObjectMeta:{webserver-deployment-69b7448995-twl7x webserver-deployment-69b7448995- deployment-9570 c76161c2-f1db-453d-906e-584aac8384eb 12392 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc003249ab7 0xc003249ab8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-lwt4b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lwt4b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.339: INFO: Pod "webserver-deployment-69b7448995-w8bt7" is not available: + &Pod{ObjectMeta:{webserver-deployment-69b7448995-w8bt7 webserver-deployment-69b7448995- deployment-9570 cb43181e-10f2-4c54-b26d-86335d288017 12281 0 2023-02-25 09:59:45 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc003249ca7 0xc003249ca8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:45 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:45 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-sdd2k,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sdd2k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:,StartTime:2023-02-25 09:59:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.340: INFO: Pod "webserver-deployment-69b7448995-xn2vh" is not available: + &Pod{ObjectMeta:{webserver-deployment-69b7448995-xn2vh webserver-deployment-69b7448995- deployment-9570 b991992b-41f1-41e7-82a7-3711983cda06 12275 0 2023-02-25 09:59:45 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 c5eaf26a-2a6e-441f-b269-1dfd6870022c 0xc003249e97 0xc003249e98}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:45 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c5eaf26a-2a6e-441f-b269-1dfd6870022c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:45 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-d96zw,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d96zw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:45 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:,StartTime:2023-02-25 09:59:45 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.340: INFO: Pod "webserver-deployment-845c8977d9-2x6m6" is not available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-2x6m6 webserver-deployment-845c8977d9- deployment-9570 c015d3ff-14f5-48f3-a2e6-56a320e83364 12390 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003058087 0xc003058088}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-qn6gg,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qn6gg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.344: INFO: Pod "webserver-deployment-845c8977d9-6xcgk" is not available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-6xcgk webserver-deployment-845c8977d9- deployment-9570 4c62de82-5650-4f82-bac3-94bee3e82f03 12393 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc0030581f0 0xc0030581f1}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-hsh58,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hsh58,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.351: INFO: Pod "webserver-deployment-845c8977d9-77gk6" is not available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-77gk6 webserver-deployment-845c8977d9- deployment-9570 874fde19-74ec-4e1f-a483-29cd4d9ffa64 12403 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc0030583b7 0xc0030583b8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-mcxqm,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-mcxqm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.352: INFO: Pod "webserver-deployment-845c8977d9-96wtz" is not available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-96wtz webserver-deployment-845c8977d9- deployment-9570 59bb96ec-53a4-4719-9f1b-204b7fbe6c36 12349 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003058587 0xc003058588}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-cwd2k,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cwd2k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.353: INFO: Pod "webserver-deployment-845c8977d9-9chbw" is not available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-9chbw webserver-deployment-845c8977d9- deployment-9570 92da79a1-6a77-400d-b52f-739dfb40a7e6 12335 0 2023-02-25 09:59:47 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003058757 0xc003058758}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:47 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:47 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-dvv52,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dvv52,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:47 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:47 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:47 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:47 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:47 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.353: INFO: Pod "webserver-deployment-845c8977d9-bd6lz" is available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-bd6lz webserver-deployment-845c8977d9- deployment-9570 6bee04e4-f18a-4ecb-928c-eab449f58ff5 12229 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003058927 0xc003058928}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.224\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-qzmmg,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qzmmg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:10.233.64.224,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://4abec4088dd7f45a8d07741449bbdcda323b1b104a6b3e3aac612214ab9f254c,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.224,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.354: INFO: Pod "webserver-deployment-845c8977d9-c9jcc" is available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-c9jcc webserver-deployment-845c8977d9- deployment-9570 51d46c16-41d4-4f5e-aa4e-e85e0c2a6c62 12235 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003058b17 0xc003058b18}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.83\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-wxzkx,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wxzkx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:10.233.65.83,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://1f1ea883ede1a3b7d04ef6d6dc15f9d91f5a23dbb37776eee35e4a760043eb87,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.83,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.354: INFO: Pod "webserver-deployment-845c8977d9-ck8xn" is available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-ck8xn webserver-deployment-845c8977d9- deployment-9570 fd954b1e-d4a1-492f-adcd-64d84826337b 12238 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003058d07 0xc003058d08}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.205\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-46zhh,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-46zhh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:10.233.65.205,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://bc982f63b06bb0a1a8bf7f3510ed2ce500c10b9875930d33142ca046f1e71570,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.205,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.354: INFO: Pod "webserver-deployment-845c8977d9-ckmhn" is available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-ckmhn webserver-deployment-845c8977d9- deployment-9570 981613e0-a706-4035-96d1-c7b6d3f29506 12227 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003058ef7 0xc003058ef8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.95\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-brb9t,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-brb9t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:10.233.64.95,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://3b602164883b7305eb49e1a195847c927a0300e6347a7798018c6159099ae3d5,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.95,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.355: INFO: Pod "webserver-deployment-845c8977d9-d5cw8" is available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-d5cw8 webserver-deployment-845c8977d9- deployment-9570 ce667004-92a2-4a94-a9cf-1e5b17be0de0 12226 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc0030590e7 0xc0030590e8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.96\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-pxf44,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pxf44,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:10.233.64.96,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://0a98f0f392f519251ccaf685893f1be5984853bad78de06700f10d1bebbc271d,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.96,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.360: INFO: Pod "webserver-deployment-845c8977d9-gn54c" is available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-gn54c webserver-deployment-845c8977d9- deployment-9570 aa3df65d-eca4-491d-8dcf-0bf6c0ddc6e7 12230 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc0030592d7 0xc0030592d8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.61\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-xztft,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xztft,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:10.233.65.61,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://9fcd0df0077547e16d67ec3c975044bd298aafcac9c1c94948c3dacba9048325,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.61,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.360: INFO: Pod "webserver-deployment-845c8977d9-jnkp2" is available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-jnkp2 webserver-deployment-845c8977d9- deployment-9570 56ed147d-2f8e-423d-914c-a80ce1c8f249 12219 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc0030594c7 0xc0030594c8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.142\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-5w675,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5w675,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:10.233.66.142,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://0af54bf851c09e577d02ea22cbe53a016306dfb7681b65a563bc80325acd505d,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.142,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.360: INFO: Pod "webserver-deployment-845c8977d9-kww24" is not available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-kww24 webserver-deployment-845c8977d9- deployment-9570 39d5cff6-4082-4d68-b609-d7e39e3a90b4 12457 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc0030596b7 0xc0030596b8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:50 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-vqbtf,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vqbtf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.361: INFO: Pod "webserver-deployment-845c8977d9-lqzcl" is available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-lqzcl webserver-deployment-845c8977d9- deployment-9570 4678a213-a27b-47b1-b043-d0f9cd25c3d8 12222 0 2023-02-25 09:59:41 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003059887 0xc003059888}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:41 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:44 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.61\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-vnkmv,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vnkmv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:44 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:41 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:10.233.66.61,StartTime:2023-02-25 09:59:41 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 09:59:43 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://48cde622b9e168bd90d228dc98cc4bc490c4dd6bae5163712e77077fbdde531f,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.61,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.361: INFO: Pod "webserver-deployment-845c8977d9-m6q4v" is not available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-m6q4v webserver-deployment-845c8977d9- deployment-9570 049cb947-a0ad-47f7-9c1f-52efcc5fd328 12391 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003059be7 0xc003059be8}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-796w5,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-796w5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.361: INFO: Pod "webserver-deployment-845c8977d9-mvmv2" is not available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-mvmv2 webserver-deployment-845c8977d9- deployment-9570 ef03a776-e514-4af1-be4d-01e53f4b6c75 12419 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003059d50 0xc003059d51}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-kkx7b,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kkx7b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.362: INFO: Pod "webserver-deployment-845c8977d9-prvk8" is not available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-prvk8 webserver-deployment-845c8977d9- deployment-9570 ebfb6305-1196-4bd8-8e2f-314935e3b098 12420 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc003059f17 0xc003059f18}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-4pwz4,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4pwz4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.363: INFO: Pod "webserver-deployment-845c8977d9-tsjd2" is not available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-tsjd2 webserver-deployment-845c8977d9- deployment-9570 776e9a27-7e3d-4631-ab1a-774d98c49743 12342 0 2023-02-25 09:59:47 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc00285e377 0xc00285e378}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:47 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-r55kc,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r55kc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.363: INFO: Pod "webserver-deployment-845c8977d9-vbf2v" is not available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-vbf2v webserver-deployment-845c8977d9- deployment-9570 b81c118c-0c53-4002-8500-983df02a5200 12444 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc00285e667 0xc00285e668}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:49 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-pq2cf,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pq2cf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 09:59:50.364: INFO: Pod "webserver-deployment-845c8977d9-xfdc9" is not available: + &Pod{ObjectMeta:{webserver-deployment-845c8977d9-xfdc9 webserver-deployment-845c8977d9- deployment-9570 aae3f967-2d08-404c-b3cd-eac3361c80af 12400 0 2023-02-25 09:59:48 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 629b2891-0215-442f-820d-aa24812ead01 0xc00285ec67 0xc00285ec68}] [] [{kube-controller-manager Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"629b2891-0215-442f-820d-aa24812ead01\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 09:59:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-gbmkp,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gbmkp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 09:59:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:,StartTime:2023-02-25 09:59:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + [AfterEach] [sig-apps] Deployment + test/e2e/framework/framework.go:187 + Feb 25 09:59:50.364: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "deployment-9570" for this suite. 02/25/23 09:59:50.463 + << End Captured GinkgoWriter Output +------------------------------ +SS +------------------------------ +[sig-storage] EmptyDir volumes + should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:216 +[BeforeEach] [sig-storage] EmptyDir volumes + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 09:59:50.507 +Feb 25 09:59:50.507: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir 02/25/23 09:59:50.514 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:59:50.596 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:59:50.605 +[It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:216 +STEP: Creating a pod to test emptydir 0777 on node default medium 02/25/23 09:59:50.624 +Feb 25 09:59:50.665: INFO: Waiting up to 5m0s for pod "pod-80c0cd56-396a-4116-afa5-903c5c674dc9" in namespace "emptydir-7203" to be "Succeeded or Failed" +Feb 25 09:59:50.688: INFO: Pod "pod-80c0cd56-396a-4116-afa5-903c5c674dc9": Phase="Pending", Reason="", readiness=false. Elapsed: 22.811442ms +Feb 25 09:59:52.696: INFO: Pod "pod-80c0cd56-396a-4116-afa5-903c5c674dc9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030550765s +Feb 25 09:59:54.699: INFO: Pod "pod-80c0cd56-396a-4116-afa5-903c5c674dc9": Phase="Pending", Reason="", readiness=false. Elapsed: 4.033339117s +Feb 25 09:59:56.713: INFO: Pod "pod-80c0cd56-396a-4116-afa5-903c5c674dc9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.047718724s +STEP: Saw pod success 02/25/23 09:59:56.713 +Feb 25 09:59:56.714: INFO: Pod "pod-80c0cd56-396a-4116-afa5-903c5c674dc9" satisfied condition "Succeeded or Failed" +Feb 25 09:59:56.723: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-80c0cd56-396a-4116-afa5-903c5c674dc9 container test-container: +STEP: delete the pod 02/25/23 09:59:56.764 +Feb 25 09:59:57.264: INFO: Waiting for pod pod-80c0cd56-396a-4116-afa5-903c5c674dc9 to disappear +Feb 25 09:59:57.497: INFO: Pod pod-80c0cd56-396a-4116-afa5-903c5c674dc9 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes + test/e2e/framework/framework.go:187 +Feb 25 09:59:57.498: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-7203" for this suite. 02/25/23 09:59:57.509 +{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","completed":48,"skipped":1077,"failed":0} +------------------------------ +• [SLOW TEST] [7.214 seconds] +[sig-storage] EmptyDir volumes +test/e2e/common/storage/framework.go:23 + should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:216 -Dec 14 15:57:35.134: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 create -f -' -Dec 14 15:57:35.922: INFO: stderr: "" -Dec 14 15:57:35.922: INFO: stdout: "deployment.apps/agnhost-replica created\n" -STEP: validating guestbook app 12/14/22 15:57:35.922 -Dec 14 15:57:35.922: INFO: Waiting for all frontend pods to be Running. -Dec 14 15:57:40.973: INFO: Waiting for frontend to serve content. -Dec 14 15:57:41.001: INFO: Trying to add a new entry to the guestbook. -Dec 14 15:57:41.022: INFO: Verifying that added entry can be retrieved. -STEP: using delete to clean up resources 12/14/22 15:57:41.039 -Dec 14 15:57:41.040: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 delete --grace-period=0 --force -f -' -Dec 14 15:57:41.198: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Dec 14 15:57:41.198: INFO: stdout: "service \"agnhost-replica\" force deleted\n" -STEP: using delete to clean up resources 12/14/22 15:57:41.198 -Dec 14 15:57:41.199: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 delete --grace-period=0 --force -f -' -Dec 14 15:57:41.368: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Dec 14 15:57:41.368: INFO: stdout: "service \"agnhost-primary\" force deleted\n" -STEP: using delete to clean up resources 12/14/22 15:57:41.368 -Dec 14 15:57:41.369: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 delete --grace-period=0 --force -f -' -Dec 14 15:57:41.540: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Dec 14 15:57:41.540: INFO: stdout: "service \"frontend\" force deleted\n" -STEP: using delete to clean up resources 12/14/22 15:57:41.54 -Dec 14 15:57:41.541: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 delete --grace-period=0 --force -f -' -Dec 14 15:57:41.690: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Dec 14 15:57:41.690: INFO: stdout: "deployment.apps \"frontend\" force deleted\n" -STEP: using delete to clean up resources 12/14/22 15:57:41.69 -Dec 14 15:57:41.691: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 delete --grace-period=0 --force -f -' -Dec 14 15:57:41.901: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Dec 14 15:57:41.901: INFO: stdout: "deployment.apps \"agnhost-primary\" force deleted\n" -STEP: using delete to clean up resources 12/14/22 15:57:41.902 -Dec 14 15:57:41.902: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 delete --grace-period=0 --force -f -' -Dec 14 15:57:42.177: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Dec 14 15:57:42.177: INFO: stdout: "deployment.apps \"agnhost-replica\" force deleted\n" -[AfterEach] [sig-cli] Kubectl client + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-storage] EmptyDir volumes + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 09:59:50.507 + Feb 25 09:59:50.507: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir 02/25/23 09:59:50.514 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:59:50.596 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:59:50.605 + [It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:216 + STEP: Creating a pod to test emptydir 0777 on node default medium 02/25/23 09:59:50.624 + Feb 25 09:59:50.665: INFO: Waiting up to 5m0s for pod "pod-80c0cd56-396a-4116-afa5-903c5c674dc9" in namespace "emptydir-7203" to be "Succeeded or Failed" + Feb 25 09:59:50.688: INFO: Pod "pod-80c0cd56-396a-4116-afa5-903c5c674dc9": Phase="Pending", Reason="", readiness=false. Elapsed: 22.811442ms + Feb 25 09:59:52.696: INFO: Pod "pod-80c0cd56-396a-4116-afa5-903c5c674dc9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030550765s + Feb 25 09:59:54.699: INFO: Pod "pod-80c0cd56-396a-4116-afa5-903c5c674dc9": Phase="Pending", Reason="", readiness=false. Elapsed: 4.033339117s + Feb 25 09:59:56.713: INFO: Pod "pod-80c0cd56-396a-4116-afa5-903c5c674dc9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.047718724s + STEP: Saw pod success 02/25/23 09:59:56.713 + Feb 25 09:59:56.714: INFO: Pod "pod-80c0cd56-396a-4116-afa5-903c5c674dc9" satisfied condition "Succeeded or Failed" + Feb 25 09:59:56.723: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-80c0cd56-396a-4116-afa5-903c5c674dc9 container test-container: + STEP: delete the pod 02/25/23 09:59:56.764 + Feb 25 09:59:57.264: INFO: Waiting for pod pod-80c0cd56-396a-4116-afa5-903c5c674dc9 to disappear + Feb 25 09:59:57.497: INFO: Pod pod-80c0cd56-396a-4116-afa5-903c5c674dc9 no longer exists + [AfterEach] [sig-storage] EmptyDir volumes + test/e2e/framework/framework.go:187 + Feb 25 09:59:57.498: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-7203" for this suite. 02/25/23 09:59:57.509 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] Projected configMap + should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:56 +[BeforeEach] [sig-storage] Projected configMap + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 09:59:57.731 +Feb 25 09:59:57.732: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 09:59:57.758 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:59:57.822 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:59:57.829 +[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:56 +STEP: Creating configMap with name projected-configmap-test-volume-78d749cd-ebb7-4a72-ab03-45221226e0b8 02/25/23 09:59:57.833 +STEP: Creating a pod to test consume configMaps 02/25/23 09:59:57.91 +Feb 25 09:59:57.929: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc" in namespace "projected-8782" to be "Succeeded or Failed" +Feb 25 09:59:57.962: INFO: Pod "pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc": Phase="Pending", Reason="", readiness=false. Elapsed: 32.38704ms +Feb 25 09:59:59.969: INFO: Pod "pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.039317392s +Feb 25 10:00:01.970: INFO: Pod "pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.040940499s +STEP: Saw pod success 02/25/23 10:00:01.97 +Feb 25 10:00:01.971: INFO: Pod "pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc" satisfied condition "Succeeded or Failed" +Feb 25 10:00:01.975: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc container agnhost-container: +STEP: delete the pod 02/25/23 10:00:01.99 +Feb 25 10:00:02.015: INFO: Waiting for pod pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc to disappear +Feb 25 10:00:02.019: INFO: Pod pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc no longer exists +[AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 -Dec 14 15:57:42.177: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-3876" for this suite. 12/14/22 15:57:42.19 -{"msg":"PASSED [sig-cli] Kubectl client Guestbook application should create and stop a working application [Conformance]","completed":46,"skipped":822,"failed":0} +Feb 25 10:00:02.020: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-8782" for this suite. 02/25/23 10:00:02.027 +{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","completed":49,"skipped":1105,"failed":0} ------------------------------ -• [SLOW TEST] [11.566 seconds] -[sig-cli] Kubectl client -test/e2e/kubectl/framework.go:23 - Guestbook application - test/e2e/kubectl/kubectl.go:367 - should create and stop a working application [Conformance] - test/e2e/kubectl/kubectl.go:392 +• [4.307 seconds] +[sig-storage] Projected configMap +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:56 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-cli] Kubectl client + [BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:57:30.644 - Dec 14 15:57:30.644: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 15:57:30.646 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:57:30.694 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:57:30.698 - [BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 - [It] should create and stop a working application [Conformance] - test/e2e/kubectl/kubectl.go:392 - STEP: creating all guestbook components 12/14/22 15:57:30.706 - Dec 14 15:57:30.707: INFO: apiVersion: v1 - kind: Service - metadata: - name: agnhost-replica - labels: - app: agnhost - role: replica - tier: backend - spec: - ports: - - port: 6379 - selector: - app: agnhost - role: replica - tier: backend - - Dec 14 15:57:30.707: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 create -f -' - Dec 14 15:57:32.359: INFO: stderr: "" - Dec 14 15:57:32.359: INFO: stdout: "service/agnhost-replica created\n" - Dec 14 15:57:32.359: INFO: apiVersion: v1 - kind: Service - metadata: - name: agnhost-primary - labels: - app: agnhost - role: primary - tier: backend - spec: - ports: - - port: 6379 - targetPort: 6379 - selector: - app: agnhost - role: primary - tier: backend - - Dec 14 15:57:32.360: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 create -f -' - Dec 14 15:57:33.973: INFO: stderr: "" - Dec 14 15:57:33.973: INFO: stdout: "service/agnhost-primary created\n" - Dec 14 15:57:33.981: INFO: apiVersion: v1 - kind: Service - metadata: - name: frontend - labels: - app: guestbook - tier: frontend - spec: - # if your cluster supports it, uncomment the following to automatically create - # an external load-balanced IP for the frontend service. - # type: LoadBalancer - ports: - - port: 80 - selector: - app: guestbook - tier: frontend - - Dec 14 15:57:33.982: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 create -f -' - Dec 14 15:57:34.427: INFO: stderr: "" - Dec 14 15:57:34.428: INFO: stdout: "service/frontend created\n" - Dec 14 15:57:34.428: INFO: apiVersion: apps/v1 - kind: Deployment - metadata: - name: frontend - spec: - replicas: 3 - selector: - matchLabels: - app: guestbook - tier: frontend - template: - metadata: - labels: - app: guestbook - tier: frontend - spec: - containers: - - name: guestbook-frontend - image: registry.k8s.io/e2e-test-images/agnhost:2.40 - args: [ "guestbook", "--backend-port", "6379" ] - resources: - requests: - cpu: 100m - memory: 100Mi - ports: - - containerPort: 80 - - Dec 14 15:57:34.431: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 create -f -' - Dec 14 15:57:34.774: INFO: stderr: "" - Dec 14 15:57:34.774: INFO: stdout: "deployment.apps/frontend created\n" - Dec 14 15:57:34.777: INFO: apiVersion: apps/v1 - kind: Deployment - metadata: - name: agnhost-primary - spec: - replicas: 1 - selector: - matchLabels: - app: agnhost - role: primary - tier: backend - template: - metadata: - labels: - app: agnhost - role: primary - tier: backend - spec: - containers: - - name: primary - image: registry.k8s.io/e2e-test-images/agnhost:2.40 - args: [ "guestbook", "--http-port", "6379" ] - resources: - requests: - cpu: 100m - memory: 100Mi - ports: - - containerPort: 6379 - - Dec 14 15:57:34.778: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 create -f -' - Dec 14 15:57:35.134: INFO: stderr: "" - Dec 14 15:57:35.134: INFO: stdout: "deployment.apps/agnhost-primary created\n" - Dec 14 15:57:35.134: INFO: apiVersion: apps/v1 - kind: Deployment - metadata: - name: agnhost-replica - spec: - replicas: 2 - selector: - matchLabels: - app: agnhost - role: replica - tier: backend - template: - metadata: - labels: - app: agnhost - role: replica - tier: backend - spec: - containers: - - name: replica - image: registry.k8s.io/e2e-test-images/agnhost:2.40 - args: [ "guestbook", "--replicaof", "agnhost-primary", "--http-port", "6379" ] - resources: - requests: - cpu: 100m - memory: 100Mi - ports: - - containerPort: 6379 - - Dec 14 15:57:35.134: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 create -f -' - Dec 14 15:57:35.922: INFO: stderr: "" - Dec 14 15:57:35.922: INFO: stdout: "deployment.apps/agnhost-replica created\n" - STEP: validating guestbook app 12/14/22 15:57:35.922 - Dec 14 15:57:35.922: INFO: Waiting for all frontend pods to be Running. - Dec 14 15:57:40.973: INFO: Waiting for frontend to serve content. - Dec 14 15:57:41.001: INFO: Trying to add a new entry to the guestbook. - Dec 14 15:57:41.022: INFO: Verifying that added entry can be retrieved. - STEP: using delete to clean up resources 12/14/22 15:57:41.039 - Dec 14 15:57:41.040: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 delete --grace-period=0 --force -f -' - Dec 14 15:57:41.198: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" - Dec 14 15:57:41.198: INFO: stdout: "service \"agnhost-replica\" force deleted\n" - STEP: using delete to clean up resources 12/14/22 15:57:41.198 - Dec 14 15:57:41.199: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 delete --grace-period=0 --force -f -' - Dec 14 15:57:41.368: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" - Dec 14 15:57:41.368: INFO: stdout: "service \"agnhost-primary\" force deleted\n" - STEP: using delete to clean up resources 12/14/22 15:57:41.368 - Dec 14 15:57:41.369: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 delete --grace-period=0 --force -f -' - Dec 14 15:57:41.540: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" - Dec 14 15:57:41.540: INFO: stdout: "service \"frontend\" force deleted\n" - STEP: using delete to clean up resources 12/14/22 15:57:41.54 - Dec 14 15:57:41.541: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 delete --grace-period=0 --force -f -' - Dec 14 15:57:41.690: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" - Dec 14 15:57:41.690: INFO: stdout: "deployment.apps \"frontend\" force deleted\n" - STEP: using delete to clean up resources 12/14/22 15:57:41.69 - Dec 14 15:57:41.691: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 delete --grace-period=0 --force -f -' - Dec 14 15:57:41.901: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" - Dec 14 15:57:41.901: INFO: stdout: "deployment.apps \"agnhost-primary\" force deleted\n" - STEP: using delete to clean up resources 12/14/22 15:57:41.902 - Dec 14 15:57:41.902: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3876 delete --grace-period=0 --force -f -' - Dec 14 15:57:42.177: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" - Dec 14 15:57:42.177: INFO: stdout: "deployment.apps \"agnhost-replica\" force deleted\n" - [AfterEach] [sig-cli] Kubectl client - test/e2e/framework/framework.go:187 - Dec 14 15:57:42.177: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-3876" for this suite. 12/14/22 15:57:42.19 - << End Captured GinkgoWriter Output ------------------------------- -[sig-apps] ReplicationController - should test the lifecycle of a ReplicationController [Conformance] - test/e2e/apps/rc.go:109 -[BeforeEach] [sig-apps] ReplicationController - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:57:42.214 -Dec 14 15:57:42.215: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename replication-controller 12/14/22 15:57:42.217 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:57:42.265 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:57:42.269 -[BeforeEach] [sig-apps] ReplicationController - test/e2e/apps/rc.go:56 -[It] should test the lifecycle of a ReplicationController [Conformance] - test/e2e/apps/rc.go:109 -STEP: creating a ReplicationController 12/14/22 15:57:42.283 -STEP: waiting for RC to be added 12/14/22 15:57:42.299 -STEP: waiting for available Replicas 12/14/22 15:57:42.299 -STEP: patching ReplicationController 12/14/22 15:57:43.346 -STEP: waiting for RC to be modified 12/14/22 15:57:43.357 -STEP: patching ReplicationController status 12/14/22 15:57:43.359 -STEP: waiting for RC to be modified 12/14/22 15:57:43.368 -STEP: waiting for available Replicas 12/14/22 15:57:43.37 -STEP: fetching ReplicationController status 12/14/22 15:57:43.379 -STEP: patching ReplicationController scale 12/14/22 15:57:43.387 -STEP: waiting for RC to be modified 12/14/22 15:57:43.42 -STEP: waiting for ReplicationController's scale to be the max amount 12/14/22 15:57:43.42 -STEP: fetching ReplicationController; ensuring that it's patched 12/14/22 15:57:44.641 -STEP: updating ReplicationController status 12/14/22 15:57:44.647 -STEP: waiting for RC to be modified 12/14/22 15:57:44.664 -STEP: listing all ReplicationControllers 12/14/22 15:57:44.664 -STEP: checking that ReplicationController has expected values 12/14/22 15:57:44.669 -STEP: deleting ReplicationControllers by collection 12/14/22 15:57:44.67 -STEP: waiting for ReplicationController to have a DELETED watchEvent 12/14/22 15:57:44.689 -[AfterEach] [sig-apps] ReplicationController - test/e2e/framework/framework.go:187 -Dec 14 15:57:44.722: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "replication-controller-5310" for this suite. 12/14/22 15:57:44.73 -{"msg":"PASSED [sig-apps] ReplicationController should test the lifecycle of a ReplicationController [Conformance]","completed":47,"skipped":822,"failed":0} ------------------------------- -• [2.524 seconds] -[sig-apps] ReplicationController -test/e2e/apps/framework.go:23 - should test the lifecycle of a ReplicationController [Conformance] - test/e2e/apps/rc.go:109 - - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] ReplicationController - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:57:42.214 - Dec 14 15:57:42.215: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename replication-controller 12/14/22 15:57:42.217 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:57:42.265 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:57:42.269 - [BeforeEach] [sig-apps] ReplicationController - test/e2e/apps/rc.go:56 - [It] should test the lifecycle of a ReplicationController [Conformance] - test/e2e/apps/rc.go:109 - STEP: creating a ReplicationController 12/14/22 15:57:42.283 - STEP: waiting for RC to be added 12/14/22 15:57:42.299 - STEP: waiting for available Replicas 12/14/22 15:57:42.299 - STEP: patching ReplicationController 12/14/22 15:57:43.346 - STEP: waiting for RC to be modified 12/14/22 15:57:43.357 - STEP: patching ReplicationController status 12/14/22 15:57:43.359 - STEP: waiting for RC to be modified 12/14/22 15:57:43.368 - STEP: waiting for available Replicas 12/14/22 15:57:43.37 - STEP: fetching ReplicationController status 12/14/22 15:57:43.379 - STEP: patching ReplicationController scale 12/14/22 15:57:43.387 - STEP: waiting for RC to be modified 12/14/22 15:57:43.42 - STEP: waiting for ReplicationController's scale to be the max amount 12/14/22 15:57:43.42 - STEP: fetching ReplicationController; ensuring that it's patched 12/14/22 15:57:44.641 - STEP: updating ReplicationController status 12/14/22 15:57:44.647 - STEP: waiting for RC to be modified 12/14/22 15:57:44.664 - STEP: listing all ReplicationControllers 12/14/22 15:57:44.664 - STEP: checking that ReplicationController has expected values 12/14/22 15:57:44.669 - STEP: deleting ReplicationControllers by collection 12/14/22 15:57:44.67 - STEP: waiting for ReplicationController to have a DELETED watchEvent 12/14/22 15:57:44.689 - [AfterEach] [sig-apps] ReplicationController - test/e2e/framework/framework.go:187 - Dec 14 15:57:44.722: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "replication-controller-5310" for this suite. 12/14/22 15:57:44.73 - << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-storage] ConfigMap - should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:56 -[BeforeEach] [sig-storage] ConfigMap - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:57:44.766 -Dec 14 15:57:44.767: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename configmap 12/14/22 15:57:44.77 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:57:44.792 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:57:44.803 -[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:56 -STEP: Creating configMap with name configmap-test-volume-d647f9c8-061a-4dc4-b8f5-866cb7a28bd2 12/14/22 15:57:44.807 -STEP: Creating a pod to test consume configMaps 12/14/22 15:57:44.821 -Dec 14 15:57:44.832: INFO: Waiting up to 5m0s for pod "pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574" in namespace "configmap-1908" to be "Succeeded or Failed" -Dec 14 15:57:44.835: INFO: Pod "pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574": Phase="Pending", Reason="", readiness=false. Elapsed: 3.148648ms -Dec 14 15:57:46.843: INFO: Pod "pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011231934s -Dec 14 15:57:48.844: INFO: Pod "pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012094449s -STEP: Saw pod success 12/14/22 15:57:48.844 -Dec 14 15:57:48.845: INFO: Pod "pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574" satisfied condition "Succeeded or Failed" -Dec 14 15:57:48.849: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574 container agnhost-container: -STEP: delete the pod 12/14/22 15:57:48.88 -Dec 14 15:57:48.906: INFO: Waiting for pod pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574 to disappear -Dec 14 15:57:48.910: INFO: Pod pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574 no longer exists -[AfterEach] [sig-storage] ConfigMap - test/e2e/framework/framework.go:187 -Dec 14 15:57:48.910: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-1908" for this suite. 12/14/22 15:57:48.916 -{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","completed":48,"skipped":851,"failed":0} ------------------------------- -• [4.159 seconds] -[sig-storage] ConfigMap -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:56 - - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] ConfigMap - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:57:44.766 - Dec 14 15:57:44.767: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename configmap 12/14/22 15:57:44.77 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:57:44.792 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:57:44.803 + STEP: Creating a kubernetes client 02/25/23 09:59:57.731 + Feb 25 09:59:57.732: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 09:59:57.758 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 09:59:57.822 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 09:59:57.829 [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:56 - STEP: Creating configMap with name configmap-test-volume-d647f9c8-061a-4dc4-b8f5-866cb7a28bd2 12/14/22 15:57:44.807 - STEP: Creating a pod to test consume configMaps 12/14/22 15:57:44.821 - Dec 14 15:57:44.832: INFO: Waiting up to 5m0s for pod "pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574" in namespace "configmap-1908" to be "Succeeded or Failed" - Dec 14 15:57:44.835: INFO: Pod "pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574": Phase="Pending", Reason="", readiness=false. Elapsed: 3.148648ms - Dec 14 15:57:46.843: INFO: Pod "pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011231934s - Dec 14 15:57:48.844: INFO: Pod "pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012094449s - STEP: Saw pod success 12/14/22 15:57:48.844 - Dec 14 15:57:48.845: INFO: Pod "pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574" satisfied condition "Succeeded or Failed" - Dec 14 15:57:48.849: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574 container agnhost-container: - STEP: delete the pod 12/14/22 15:57:48.88 - Dec 14 15:57:48.906: INFO: Waiting for pod pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574 to disappear - Dec 14 15:57:48.910: INFO: Pod pod-configmaps-a8f68960-f296-4376-a370-b9a3935a4574 no longer exists - [AfterEach] [sig-storage] ConfigMap + test/e2e/common/storage/projected_configmap.go:56 + STEP: Creating configMap with name projected-configmap-test-volume-78d749cd-ebb7-4a72-ab03-45221226e0b8 02/25/23 09:59:57.833 + STEP: Creating a pod to test consume configMaps 02/25/23 09:59:57.91 + Feb 25 09:59:57.929: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc" in namespace "projected-8782" to be "Succeeded or Failed" + Feb 25 09:59:57.962: INFO: Pod "pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc": Phase="Pending", Reason="", readiness=false. Elapsed: 32.38704ms + Feb 25 09:59:59.969: INFO: Pod "pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.039317392s + Feb 25 10:00:01.970: INFO: Pod "pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.040940499s + STEP: Saw pod success 02/25/23 10:00:01.97 + Feb 25 10:00:01.971: INFO: Pod "pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc" satisfied condition "Succeeded or Failed" + Feb 25 10:00:01.975: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc container agnhost-container: + STEP: delete the pod 02/25/23 10:00:01.99 + Feb 25 10:00:02.015: INFO: Waiting for pod pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc to disappear + Feb 25 10:00:02.019: INFO: Pod pod-projected-configmaps-bf4cf668-eb3d-4020-93a8-48358c23e5dc no longer exists + [AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 - Dec 14 15:57:48.910: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "configmap-1908" for this suite. 12/14/22 15:57:48.916 + Feb 25 10:00:02.020: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-8782" for this suite. 02/25/23 10:00:02.027 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSS +S ------------------------------ [sig-network] Services - should test the lifecycle of an Endpoint [Conformance] - test/e2e/network/service.go:3231 + should be able to change the type from ExternalName to NodePort [Conformance] + test/e2e/network/service.go:1443 [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:57:48.928 -Dec 14 15:57:48.928: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 15:57:48.931 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:57:48.95 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:57:48.953 +STEP: Creating a kubernetes client 02/25/23 10:00:02.041 +Feb 25 10:00:02.042: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 10:00:02.045 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:00:02.07 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:00:02.075 [BeforeEach] [sig-network] Services test/e2e/network/service.go:758 -[It] should test the lifecycle of an Endpoint [Conformance] - test/e2e/network/service.go:3231 -STEP: creating an Endpoint 12/14/22 15:57:48.961 -STEP: waiting for available Endpoint 12/14/22 15:57:48.968 -STEP: listing all Endpoints 12/14/22 15:57:48.971 -STEP: updating the Endpoint 12/14/22 15:57:48.983 -STEP: fetching the Endpoint 12/14/22 15:57:48.993 -STEP: patching the Endpoint 12/14/22 15:57:48.997 -STEP: fetching the Endpoint 12/14/22 15:57:49.009 -STEP: deleting the Endpoint by Collection 12/14/22 15:57:49.012 -STEP: waiting for Endpoint deletion 12/14/22 15:57:49.025 -STEP: fetching the Endpoint 12/14/22 15:57:49.027 +[It] should be able to change the type from ExternalName to NodePort [Conformance] + test/e2e/network/service.go:1443 +STEP: creating a service externalname-service with the type=ExternalName in namespace services-5413 02/25/23 10:00:02.081 +STEP: changing the ExternalName service to type=NodePort 02/25/23 10:00:02.09 +STEP: creating replication controller externalname-service in namespace services-5413 02/25/23 10:00:02.13 +I0225 10:00:02.145238 13 runners.go:193] Created replication controller with name: externalname-service, namespace: services-5413, replica count: 2 +I0225 10:00:05.199040 13 runners.go:193] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Feb 25 10:00:05.199: INFO: Creating new exec pod +Feb 25 10:00:05.208: INFO: Waiting up to 5m0s for pod "execpod46qmz" in namespace "services-5413" to be "running" +Feb 25 10:00:05.216: INFO: Pod "execpod46qmz": Phase="Pending", Reason="", readiness=false. Elapsed: 7.251871ms +Feb 25 10:00:07.233: INFO: Pod "execpod46qmz": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024066991s +Feb 25 10:00:09.225: INFO: Pod "execpod46qmz": Phase="Running", Reason="", readiness=true. Elapsed: 4.016192716s +Feb 25 10:00:09.225: INFO: Pod "execpod46qmz" satisfied condition "running" +Feb 25 10:00:10.240: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-5413 exec execpod46qmz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 externalname-service 80' +Feb 25 10:00:10.523: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n" +Feb 25 10:00:10.523: INFO: stdout: "externalname-service-67s67" +Feb 25 10:00:10.524: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-5413 exec execpod46qmz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.47.179 80' +Feb 25 10:00:10.765: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.47.179 80\nConnection to 10.233.47.179 80 port [tcp/http] succeeded!\n" +Feb 25 10:00:10.765: INFO: stdout: "externalname-service-67s67" +Feb 25 10:00:10.766: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-5413 exec execpod46qmz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.250 30893' +Feb 25 10:00:11.035: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.250 30893\nConnection to 192.168.121.250 30893 port [tcp/*] succeeded!\n" +Feb 25 10:00:11.035: INFO: stdout: "externalname-service-q2slm" +Feb 25 10:00:11.035: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-5413 exec execpod46qmz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.117 30893' +Feb 25 10:00:11.263: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.117 30893\nConnection to 192.168.121.117 30893 port [tcp/*] succeeded!\n" +Feb 25 10:00:11.263: INFO: stdout: "externalname-service-q2slm" +Feb 25 10:00:11.263: INFO: Cleaning up the ExternalName to NodePort test service [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 15:57:49.032: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-6682" for this suite. 12/14/22 15:57:49.038 +Feb 25 10:00:11.313: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-5413" for this suite. 02/25/23 10:00:11.33 [AfterEach] [sig-network] Services test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should test the lifecycle of an Endpoint [Conformance]","completed":49,"skipped":879,"failed":0} +{"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to NodePort [Conformance]","completed":50,"skipped":1106,"failed":0} ------------------------------ -• [0.119 seconds] +• [SLOW TEST] [9.310 seconds] [sig-network] Services test/e2e/network/common/framework.go:23 - should test the lifecycle of an Endpoint [Conformance] - test/e2e/network/service.go:3231 + should be able to change the type from ExternalName to NodePort [Conformance] + test/e2e/network/service.go:1443 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:57:48.928 - Dec 14 15:57:48.928: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 15:57:48.931 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:57:48.95 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:57:48.953 + STEP: Creating a kubernetes client 02/25/23 10:00:02.041 + Feb 25 10:00:02.042: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 10:00:02.045 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:00:02.07 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:00:02.075 [BeforeEach] [sig-network] Services test/e2e/network/service.go:758 - [It] should test the lifecycle of an Endpoint [Conformance] - test/e2e/network/service.go:3231 - STEP: creating an Endpoint 12/14/22 15:57:48.961 - STEP: waiting for available Endpoint 12/14/22 15:57:48.968 - STEP: listing all Endpoints 12/14/22 15:57:48.971 - STEP: updating the Endpoint 12/14/22 15:57:48.983 - STEP: fetching the Endpoint 12/14/22 15:57:48.993 - STEP: patching the Endpoint 12/14/22 15:57:48.997 - STEP: fetching the Endpoint 12/14/22 15:57:49.009 - STEP: deleting the Endpoint by Collection 12/14/22 15:57:49.012 - STEP: waiting for Endpoint deletion 12/14/22 15:57:49.025 - STEP: fetching the Endpoint 12/14/22 15:57:49.027 + [It] should be able to change the type from ExternalName to NodePort [Conformance] + test/e2e/network/service.go:1443 + STEP: creating a service externalname-service with the type=ExternalName in namespace services-5413 02/25/23 10:00:02.081 + STEP: changing the ExternalName service to type=NodePort 02/25/23 10:00:02.09 + STEP: creating replication controller externalname-service in namespace services-5413 02/25/23 10:00:02.13 + I0225 10:00:02.145238 13 runners.go:193] Created replication controller with name: externalname-service, namespace: services-5413, replica count: 2 + I0225 10:00:05.199040 13 runners.go:193] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + Feb 25 10:00:05.199: INFO: Creating new exec pod + Feb 25 10:00:05.208: INFO: Waiting up to 5m0s for pod "execpod46qmz" in namespace "services-5413" to be "running" + Feb 25 10:00:05.216: INFO: Pod "execpod46qmz": Phase="Pending", Reason="", readiness=false. Elapsed: 7.251871ms + Feb 25 10:00:07.233: INFO: Pod "execpod46qmz": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024066991s + Feb 25 10:00:09.225: INFO: Pod "execpod46qmz": Phase="Running", Reason="", readiness=true. Elapsed: 4.016192716s + Feb 25 10:00:09.225: INFO: Pod "execpod46qmz" satisfied condition "running" + Feb 25 10:00:10.240: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-5413 exec execpod46qmz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 externalname-service 80' + Feb 25 10:00:10.523: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n" + Feb 25 10:00:10.523: INFO: stdout: "externalname-service-67s67" + Feb 25 10:00:10.524: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-5413 exec execpod46qmz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.47.179 80' + Feb 25 10:00:10.765: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.47.179 80\nConnection to 10.233.47.179 80 port [tcp/http] succeeded!\n" + Feb 25 10:00:10.765: INFO: stdout: "externalname-service-67s67" + Feb 25 10:00:10.766: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-5413 exec execpod46qmz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.250 30893' + Feb 25 10:00:11.035: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.250 30893\nConnection to 192.168.121.250 30893 port [tcp/*] succeeded!\n" + Feb 25 10:00:11.035: INFO: stdout: "externalname-service-q2slm" + Feb 25 10:00:11.035: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-5413 exec execpod46qmz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.117 30893' + Feb 25 10:00:11.263: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.117 30893\nConnection to 192.168.121.117 30893 port [tcp/*] succeeded!\n" + Feb 25 10:00:11.263: INFO: stdout: "externalname-service-q2slm" + Feb 25 10:00:11.263: INFO: Cleaning up the ExternalName to NodePort test service [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 15:57:49.032: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-6682" for this suite. 12/14/22 15:57:49.038 + Feb 25 10:00:11.313: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-5413" for this suite. 02/25/23 10:00:11.33 [AfterEach] [sig-network] Services test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSS ------------------------------ -[sig-apps] ReplicaSet - should serve a basic image on each replica with a public image [Conformance] - test/e2e/apps/replica_set.go:111 -[BeforeEach] [sig-apps] ReplicaSet +[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] + should validate Statefulset Status endpoints [Conformance] + test/e2e/apps/statefulset.go:975 +[BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:57:49.052 -Dec 14 15:57:49.052: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename replicaset 12/14/22 15:57:49.054 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:57:49.091 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:57:49.095 -[It] should serve a basic image on each replica with a public image [Conformance] - test/e2e/apps/replica_set.go:111 -Dec 14 15:57:49.099: INFO: Creating ReplicaSet my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d -Dec 14 15:57:49.133: INFO: Pod name my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d: Found 0 pods out of 1 -Dec 14 15:57:54.143: INFO: Pod name my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d: Found 1 pods out of 1 -Dec 14 15:57:54.144: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d" is running -Dec 14 15:57:54.144: INFO: Waiting up to 5m0s for pod "my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d-hj6vv" in namespace "replicaset-3074" to be "running" -Dec 14 15:57:54.148: INFO: Pod "my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d-hj6vv": Phase="Running", Reason="", readiness=true. Elapsed: 3.974678ms -Dec 14 15:57:54.148: INFO: Pod "my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d-hj6vv" satisfied condition "running" -Dec 14 15:57:54.148: INFO: Pod "my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d-hj6vv" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 15:57:49 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 15:57:50 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 15:57:50 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 15:57:49 +0000 UTC Reason: Message:}]) -Dec 14 15:57:54.148: INFO: Trying to dial the pod -Dec 14 15:57:59.165: INFO: Controller my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d: Got expected result from replica 1 [my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d-hj6vv]: "my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d-hj6vv", 1 of 1 required successes so far -[AfterEach] [sig-apps] ReplicaSet +STEP: Creating a kubernetes client 02/25/23 10:00:11.353 +Feb 25 10:00:11.353: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename statefulset 02/25/23 10:00:11.357 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:00:11.413 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:00:11.417 +[BeforeEach] [sig-apps] StatefulSet + test/e2e/apps/statefulset.go:96 +[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:111 +STEP: Creating service test in namespace statefulset-5114 02/25/23 10:00:11.435 +[It] should validate Statefulset Status endpoints [Conformance] + test/e2e/apps/statefulset.go:975 +STEP: Creating statefulset ss in namespace statefulset-5114 02/25/23 10:00:11.468 +Feb 25 10:00:11.491: INFO: Found 0 stateful pods, waiting for 1 +Feb 25 10:00:21.506: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true +STEP: Patch Statefulset to include a label 02/25/23 10:00:21.519 +STEP: Getting /status 02/25/23 10:00:21.535 +Feb 25 10:00:21.548: INFO: StatefulSet ss has Conditions: []v1.StatefulSetCondition(nil) +STEP: updating the StatefulSet Status 02/25/23 10:00:21.548 +Feb 25 10:00:21.573: INFO: updatedStatus.Conditions: []v1.StatefulSetCondition{v1.StatefulSetCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} +STEP: watching for the statefulset status to be updated 02/25/23 10:00:21.573 +Feb 25 10:00:21.579: INFO: Observed &StatefulSet event: ADDED +Feb 25 10:00:21.579: INFO: Found Statefulset ss in namespace statefulset-5114 with labels: map[e2e:testing] annotations: map[] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} +Feb 25 10:00:21.579: INFO: Statefulset ss has an updated status +STEP: patching the Statefulset Status 02/25/23 10:00:21.579 +Feb 25 10:00:21.580: INFO: Patch payload: {"status":{"conditions":[{"type":"StatusPatched","status":"True"}]}} +Feb 25 10:00:21.593: INFO: Patched status conditions: []v1.StatefulSetCondition{v1.StatefulSetCondition{Type:"StatusPatched", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"", Message:""}} +STEP: watching for the Statefulset status to be patched 02/25/23 10:00:21.593 +Feb 25 10:00:21.597: INFO: Observed &StatefulSet event: ADDED +Feb 25 10:00:21.597: INFO: Observed Statefulset ss in namespace statefulset-5114 with annotations: map[] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} +Feb 25 10:00:21.598: INFO: Observed &StatefulSet event: MODIFIED +[AfterEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:122 +Feb 25 10:00:21.599: INFO: Deleting all statefulset in ns statefulset-5114 +Feb 25 10:00:21.605: INFO: Scaling statefulset ss to 0 +Feb 25 10:00:31.658: INFO: Waiting for statefulset status.replicas updated to 0 +Feb 25 10:00:31.663: INFO: Deleting statefulset ss +[AfterEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:187 -Dec 14 15:57:59.166: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "replicaset-3074" for this suite. 12/14/22 15:57:59.172 -{"msg":"PASSED [sig-apps] ReplicaSet should serve a basic image on each replica with a public image [Conformance]","completed":50,"skipped":945,"failed":0} +Feb 25 10:00:31.700: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "statefulset-5114" for this suite. 02/25/23 10:00:31.71 +{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should validate Statefulset Status endpoints [Conformance]","completed":51,"skipped":1109,"failed":0} ------------------------------ -• [SLOW TEST] [10.128 seconds] -[sig-apps] ReplicaSet +• [SLOW TEST] [20.371 seconds] +[sig-apps] StatefulSet test/e2e/apps/framework.go:23 - should serve a basic image on each replica with a public image [Conformance] - test/e2e/apps/replica_set.go:111 + Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:101 + should validate Statefulset Status endpoints [Conformance] + test/e2e/apps/statefulset.go:975 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] ReplicaSet + [BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:57:49.052 - Dec 14 15:57:49.052: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename replicaset 12/14/22 15:57:49.054 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:57:49.091 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:57:49.095 - [It] should serve a basic image on each replica with a public image [Conformance] - test/e2e/apps/replica_set.go:111 - Dec 14 15:57:49.099: INFO: Creating ReplicaSet my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d - Dec 14 15:57:49.133: INFO: Pod name my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d: Found 0 pods out of 1 - Dec 14 15:57:54.143: INFO: Pod name my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d: Found 1 pods out of 1 - Dec 14 15:57:54.144: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d" is running - Dec 14 15:57:54.144: INFO: Waiting up to 5m0s for pod "my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d-hj6vv" in namespace "replicaset-3074" to be "running" - Dec 14 15:57:54.148: INFO: Pod "my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d-hj6vv": Phase="Running", Reason="", readiness=true. Elapsed: 3.974678ms - Dec 14 15:57:54.148: INFO: Pod "my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d-hj6vv" satisfied condition "running" - Dec 14 15:57:54.148: INFO: Pod "my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d-hj6vv" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 15:57:49 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 15:57:50 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 15:57:50 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 15:57:49 +0000 UTC Reason: Message:}]) - Dec 14 15:57:54.148: INFO: Trying to dial the pod - Dec 14 15:57:59.165: INFO: Controller my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d: Got expected result from replica 1 [my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d-hj6vv]: "my-hostname-basic-bb7148f7-6b82-4af1-a4bc-6e15f8ac3d1d-hj6vv", 1 of 1 required successes so far - [AfterEach] [sig-apps] ReplicaSet + STEP: Creating a kubernetes client 02/25/23 10:00:11.353 + Feb 25 10:00:11.353: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename statefulset 02/25/23 10:00:11.357 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:00:11.413 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:00:11.417 + [BeforeEach] [sig-apps] StatefulSet + test/e2e/apps/statefulset.go:96 + [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:111 + STEP: Creating service test in namespace statefulset-5114 02/25/23 10:00:11.435 + [It] should validate Statefulset Status endpoints [Conformance] + test/e2e/apps/statefulset.go:975 + STEP: Creating statefulset ss in namespace statefulset-5114 02/25/23 10:00:11.468 + Feb 25 10:00:11.491: INFO: Found 0 stateful pods, waiting for 1 + Feb 25 10:00:21.506: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true + STEP: Patch Statefulset to include a label 02/25/23 10:00:21.519 + STEP: Getting /status 02/25/23 10:00:21.535 + Feb 25 10:00:21.548: INFO: StatefulSet ss has Conditions: []v1.StatefulSetCondition(nil) + STEP: updating the StatefulSet Status 02/25/23 10:00:21.548 + Feb 25 10:00:21.573: INFO: updatedStatus.Conditions: []v1.StatefulSetCondition{v1.StatefulSetCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} + STEP: watching for the statefulset status to be updated 02/25/23 10:00:21.573 + Feb 25 10:00:21.579: INFO: Observed &StatefulSet event: ADDED + Feb 25 10:00:21.579: INFO: Found Statefulset ss in namespace statefulset-5114 with labels: map[e2e:testing] annotations: map[] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} + Feb 25 10:00:21.579: INFO: Statefulset ss has an updated status + STEP: patching the Statefulset Status 02/25/23 10:00:21.579 + Feb 25 10:00:21.580: INFO: Patch payload: {"status":{"conditions":[{"type":"StatusPatched","status":"True"}]}} + Feb 25 10:00:21.593: INFO: Patched status conditions: []v1.StatefulSetCondition{v1.StatefulSetCondition{Type:"StatusPatched", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"", Message:""}} + STEP: watching for the Statefulset status to be patched 02/25/23 10:00:21.593 + Feb 25 10:00:21.597: INFO: Observed &StatefulSet event: ADDED + Feb 25 10:00:21.597: INFO: Observed Statefulset ss in namespace statefulset-5114 with annotations: map[] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} + Feb 25 10:00:21.598: INFO: Observed &StatefulSet event: MODIFIED + [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:122 + Feb 25 10:00:21.599: INFO: Deleting all statefulset in ns statefulset-5114 + Feb 25 10:00:21.605: INFO: Scaling statefulset ss to 0 + Feb 25 10:00:31.658: INFO: Waiting for statefulset status.replicas updated to 0 + Feb 25 10:00:31.663: INFO: Deleting statefulset ss + [AfterEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:187 - Dec 14 15:57:59.166: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "replicaset-3074" for this suite. 12/14/22 15:57:59.172 + Feb 25 10:00:31.700: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "statefulset-5114" for this suite. 02/25/23 10:00:31.71 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSS +SSSSSS ------------------------------ -[sig-api-machinery] Garbage collector - should delete pods created by rc when not orphaning [Conformance] - test/e2e/apimachinery/garbage_collector.go:312 -[BeforeEach] [sig-api-machinery] Garbage collector +[sig-node] Probing container + should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:165 +[BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:57:59.186 -Dec 14 15:57:59.186: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename gc 12/14/22 15:57:59.188 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:57:59.209 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:57:59.213 -[It] should delete pods created by rc when not orphaning [Conformance] - test/e2e/apimachinery/garbage_collector.go:312 -STEP: create the rc 12/14/22 15:57:59.222 -STEP: delete the rc 12/14/22 15:58:04.235 -STEP: wait for all pods to be garbage collected 12/14/22 15:58:04.247 -STEP: Gathering metrics 12/14/22 15:58:09.265 -Dec 14 15:58:09.313: INFO: Waiting up to 5m0s for pod "kube-controller-manager-kiedai4iakae-2" in namespace "kube-system" to be "running and ready" -Dec 14 15:58:09.318: INFO: Pod "kube-controller-manager-kiedai4iakae-2": Phase="Running", Reason="", readiness=true. Elapsed: 4.975885ms -Dec 14 15:58:09.319: INFO: The phase of Pod kube-controller-manager-kiedai4iakae-2 is Running (Ready = true) -Dec 14 15:58:09.319: INFO: Pod "kube-controller-manager-kiedai4iakae-2" satisfied condition "running and ready" -Dec 14 15:58:09.442: INFO: For apiserver_request_total: -For apiserver_request_latency_seconds: -For apiserver_init_events_total: -For garbage_collector_attempt_to_delete_queue_latency: -For garbage_collector_attempt_to_delete_work_duration: -For garbage_collector_attempt_to_orphan_queue_latency: -For garbage_collector_attempt_to_orphan_work_duration: -For garbage_collector_dirty_processing_latency_microseconds: -For garbage_collector_event_processing_latency_microseconds: -For garbage_collector_graph_changes_queue_latency: -For garbage_collector_graph_changes_work_duration: -For garbage_collector_orphan_processing_latency_microseconds: -For namespace_queue_latency: -For namespace_queue_latency_sum: -For namespace_queue_latency_count: -For namespace_retries: -For namespace_work_duration: -For namespace_work_duration_sum: -For namespace_work_duration_count: -For function_duration_seconds: -For errors_total: -For evicted_pods_total: - -[AfterEach] [sig-api-machinery] Garbage collector +STEP: Creating a kubernetes client 02/25/23 10:00:31.725 +Feb 25 10:00:31.726: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-probe 02/25/23 10:00:31.731 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:00:31.751 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:00:31.761 +[BeforeEach] [sig-node] Probing container + test/e2e/common/node/container_probe.go:59 +[It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:165 +STEP: Creating pod liveness-1b60ac0b-4438-40a8-baca-4636aff673d3 in namespace container-probe-5374 02/25/23 10:00:31.766 +Feb 25 10:00:31.780: INFO: Waiting up to 5m0s for pod "liveness-1b60ac0b-4438-40a8-baca-4636aff673d3" in namespace "container-probe-5374" to be "not pending" +Feb 25 10:00:31.787: INFO: Pod "liveness-1b60ac0b-4438-40a8-baca-4636aff673d3": Phase="Pending", Reason="", readiness=false. Elapsed: 6.354697ms +Feb 25 10:00:33.793: INFO: Pod "liveness-1b60ac0b-4438-40a8-baca-4636aff673d3": Phase="Running", Reason="", readiness=true. Elapsed: 2.012576409s +Feb 25 10:00:33.793: INFO: Pod "liveness-1b60ac0b-4438-40a8-baca-4636aff673d3" satisfied condition "not pending" +Feb 25 10:00:33.793: INFO: Started pod liveness-1b60ac0b-4438-40a8-baca-4636aff673d3 in namespace container-probe-5374 +STEP: checking the pod's current state and verifying that restartCount is present 02/25/23 10:00:33.793 +Feb 25 10:00:33.800: INFO: Initial restart count of pod liveness-1b60ac0b-4438-40a8-baca-4636aff673d3 is 0 +Feb 25 10:00:53.886: INFO: Restart count of pod container-probe-5374/liveness-1b60ac0b-4438-40a8-baca-4636aff673d3 is now 1 (20.085967368s elapsed) +STEP: deleting the pod 02/25/23 10:00:53.886 +[AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 -Dec 14 15:58:09.442: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "gc-3400" for this suite. 12/14/22 15:58:09.453 -{"msg":"PASSED [sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance]","completed":51,"skipped":961,"failed":0} +Feb 25 10:00:53.902: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-5374" for this suite. 02/25/23 10:00:53.91 +{"msg":"PASSED [sig-node] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","completed":52,"skipped":1115,"failed":0} ------------------------------ -• [SLOW TEST] [10.280 seconds] -[sig-api-machinery] Garbage collector -test/e2e/apimachinery/framework.go:23 - should delete pods created by rc when not orphaning [Conformance] - test/e2e/apimachinery/garbage_collector.go:312 +• [SLOW TEST] [22.194 seconds] +[sig-node] Probing container +test/e2e/common/node/framework.go:23 + should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:165 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Garbage collector + [BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:57:59.186 - Dec 14 15:57:59.186: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename gc 12/14/22 15:57:59.188 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:57:59.209 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:57:59.213 - [It] should delete pods created by rc when not orphaning [Conformance] - test/e2e/apimachinery/garbage_collector.go:312 - STEP: create the rc 12/14/22 15:57:59.222 - STEP: delete the rc 12/14/22 15:58:04.235 - STEP: wait for all pods to be garbage collected 12/14/22 15:58:04.247 - STEP: Gathering metrics 12/14/22 15:58:09.265 - Dec 14 15:58:09.313: INFO: Waiting up to 5m0s for pod "kube-controller-manager-kiedai4iakae-2" in namespace "kube-system" to be "running and ready" - Dec 14 15:58:09.318: INFO: Pod "kube-controller-manager-kiedai4iakae-2": Phase="Running", Reason="", readiness=true. Elapsed: 4.975885ms - Dec 14 15:58:09.319: INFO: The phase of Pod kube-controller-manager-kiedai4iakae-2 is Running (Ready = true) - Dec 14 15:58:09.319: INFO: Pod "kube-controller-manager-kiedai4iakae-2" satisfied condition "running and ready" - Dec 14 15:58:09.442: INFO: For apiserver_request_total: - For apiserver_request_latency_seconds: - For apiserver_init_events_total: - For garbage_collector_attempt_to_delete_queue_latency: - For garbage_collector_attempt_to_delete_work_duration: - For garbage_collector_attempt_to_orphan_queue_latency: - For garbage_collector_attempt_to_orphan_work_duration: - For garbage_collector_dirty_processing_latency_microseconds: - For garbage_collector_event_processing_latency_microseconds: - For garbage_collector_graph_changes_queue_latency: - For garbage_collector_graph_changes_work_duration: - For garbage_collector_orphan_processing_latency_microseconds: - For namespace_queue_latency: - For namespace_queue_latency_sum: - For namespace_queue_latency_count: - For namespace_retries: - For namespace_work_duration: - For namespace_work_duration_sum: - For namespace_work_duration_count: - For function_duration_seconds: - For errors_total: - For evicted_pods_total: - - [AfterEach] [sig-api-machinery] Garbage collector + STEP: Creating a kubernetes client 02/25/23 10:00:31.725 + Feb 25 10:00:31.726: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-probe 02/25/23 10:00:31.731 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:00:31.751 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:00:31.761 + [BeforeEach] [sig-node] Probing container + test/e2e/common/node/container_probe.go:59 + [It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:165 + STEP: Creating pod liveness-1b60ac0b-4438-40a8-baca-4636aff673d3 in namespace container-probe-5374 02/25/23 10:00:31.766 + Feb 25 10:00:31.780: INFO: Waiting up to 5m0s for pod "liveness-1b60ac0b-4438-40a8-baca-4636aff673d3" in namespace "container-probe-5374" to be "not pending" + Feb 25 10:00:31.787: INFO: Pod "liveness-1b60ac0b-4438-40a8-baca-4636aff673d3": Phase="Pending", Reason="", readiness=false. Elapsed: 6.354697ms + Feb 25 10:00:33.793: INFO: Pod "liveness-1b60ac0b-4438-40a8-baca-4636aff673d3": Phase="Running", Reason="", readiness=true. Elapsed: 2.012576409s + Feb 25 10:00:33.793: INFO: Pod "liveness-1b60ac0b-4438-40a8-baca-4636aff673d3" satisfied condition "not pending" + Feb 25 10:00:33.793: INFO: Started pod liveness-1b60ac0b-4438-40a8-baca-4636aff673d3 in namespace container-probe-5374 + STEP: checking the pod's current state and verifying that restartCount is present 02/25/23 10:00:33.793 + Feb 25 10:00:33.800: INFO: Initial restart count of pod liveness-1b60ac0b-4438-40a8-baca-4636aff673d3 is 0 + Feb 25 10:00:53.886: INFO: Restart count of pod container-probe-5374/liveness-1b60ac0b-4438-40a8-baca-4636aff673d3 is now 1 (20.085967368s elapsed) + STEP: deleting the pod 02/25/23 10:00:53.886 + [AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 - Dec 14 15:58:09.442: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "gc-3400" for this suite. 12/14/22 15:58:09.453 + Feb 25 10:00:53.902: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-probe-5374" for this suite. 02/25/23 10:00:53.91 << End Captured GinkgoWriter Output ------------------------------ -SS +SSSS ------------------------------ -[sig-apps] Job - should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] - test/e2e/apps/job.go:254 -[BeforeEach] [sig-apps] Job +[sig-storage] EmptyDir volumes + pod should support shared volumes between containers [Conformance] + test/e2e/common/storage/empty_dir.go:226 +[BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:58:09.468 -Dec 14 15:58:09.469: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename job 12/14/22 15:58:09.473 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:09.497 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:09.501 -[It] should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] - test/e2e/apps/job.go:254 -STEP: Creating a job 12/14/22 15:58:09.504 -STEP: Ensuring job reaches completions 12/14/22 15:58:09.511 -[AfterEach] [sig-apps] Job +STEP: Creating a kubernetes client 02/25/23 10:00:53.93 +Feb 25 10:00:53.930: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir 02/25/23 10:00:53.934 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:00:53.96 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:00:53.967 +[It] pod should support shared volumes between containers [Conformance] + test/e2e/common/storage/empty_dir.go:226 +STEP: Creating Pod 02/25/23 10:00:53.975 +Feb 25 10:00:53.991: INFO: Waiting up to 5m0s for pod "pod-sharedvolume-874698c9-9b9f-425d-a7b4-025716401ca7" in namespace "emptydir-7288" to be "running" +Feb 25 10:00:53.996: INFO: Pod "pod-sharedvolume-874698c9-9b9f-425d-a7b4-025716401ca7": Phase="Pending", Reason="", readiness=false. Elapsed: 5.44001ms +Feb 25 10:00:56.008: INFO: Pod "pod-sharedvolume-874698c9-9b9f-425d-a7b4-025716401ca7": Phase="Running", Reason="", readiness=false. Elapsed: 2.017275994s +Feb 25 10:00:56.008: INFO: Pod "pod-sharedvolume-874698c9-9b9f-425d-a7b4-025716401ca7" satisfied condition "running" +STEP: Reading file content from the nginx-container 02/25/23 10:00:56.009 +Feb 25 10:00:56.009: INFO: ExecWithOptions {Command:[/bin/sh -c cat /usr/share/volumeshare/shareddata.txt] Namespace:emptydir-7288 PodName:pod-sharedvolume-874698c9-9b9f-425d-a7b4-025716401ca7 ContainerName:busybox-main-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:00:56.010: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:00:56.012: INFO: ExecWithOptions: Clientset creation +Feb 25 10:00:56.012: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/emptydir-7288/pods/pod-sharedvolume-874698c9-9b9f-425d-a7b4-025716401ca7/exec?command=%2Fbin%2Fsh&command=-c&command=cat+%2Fusr%2Fshare%2Fvolumeshare%2Fshareddata.txt&container=busybox-main-container&container=busybox-main-container&stderr=true&stdout=true) +Feb 25 10:00:56.119: INFO: Exec stderr: "" +[AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 -Dec 14 15:58:21.519: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "job-605" for this suite. 12/14/22 15:58:21.526 -{"msg":"PASSED [sig-apps] Job should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]","completed":52,"skipped":963,"failed":0} +Feb 25 10:00:56.120: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-7288" for this suite. 02/25/23 10:00:56.128 +{"msg":"PASSED [sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance]","completed":53,"skipped":1119,"failed":0} ------------------------------ -• [SLOW TEST] [12.065 seconds] -[sig-apps] Job -test/e2e/apps/framework.go:23 - should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] - test/e2e/apps/job.go:254 +• [2.209 seconds] +[sig-storage] EmptyDir volumes +test/e2e/common/storage/framework.go:23 + pod should support shared volumes between containers [Conformance] + test/e2e/common/storage/empty_dir.go:226 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Job + [BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:58:09.468 - Dec 14 15:58:09.469: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename job 12/14/22 15:58:09.473 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:09.497 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:09.501 - [It] should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] - test/e2e/apps/job.go:254 - STEP: Creating a job 12/14/22 15:58:09.504 - STEP: Ensuring job reaches completions 12/14/22 15:58:09.511 - [AfterEach] [sig-apps] Job + STEP: Creating a kubernetes client 02/25/23 10:00:53.93 + Feb 25 10:00:53.930: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir 02/25/23 10:00:53.934 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:00:53.96 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:00:53.967 + [It] pod should support shared volumes between containers [Conformance] + test/e2e/common/storage/empty_dir.go:226 + STEP: Creating Pod 02/25/23 10:00:53.975 + Feb 25 10:00:53.991: INFO: Waiting up to 5m0s for pod "pod-sharedvolume-874698c9-9b9f-425d-a7b4-025716401ca7" in namespace "emptydir-7288" to be "running" + Feb 25 10:00:53.996: INFO: Pod "pod-sharedvolume-874698c9-9b9f-425d-a7b4-025716401ca7": Phase="Pending", Reason="", readiness=false. Elapsed: 5.44001ms + Feb 25 10:00:56.008: INFO: Pod "pod-sharedvolume-874698c9-9b9f-425d-a7b4-025716401ca7": Phase="Running", Reason="", readiness=false. Elapsed: 2.017275994s + Feb 25 10:00:56.008: INFO: Pod "pod-sharedvolume-874698c9-9b9f-425d-a7b4-025716401ca7" satisfied condition "running" + STEP: Reading file content from the nginx-container 02/25/23 10:00:56.009 + Feb 25 10:00:56.009: INFO: ExecWithOptions {Command:[/bin/sh -c cat /usr/share/volumeshare/shareddata.txt] Namespace:emptydir-7288 PodName:pod-sharedvolume-874698c9-9b9f-425d-a7b4-025716401ca7 ContainerName:busybox-main-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:00:56.010: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:00:56.012: INFO: ExecWithOptions: Clientset creation + Feb 25 10:00:56.012: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/emptydir-7288/pods/pod-sharedvolume-874698c9-9b9f-425d-a7b4-025716401ca7/exec?command=%2Fbin%2Fsh&command=-c&command=cat+%2Fusr%2Fshare%2Fvolumeshare%2Fshareddata.txt&container=busybox-main-container&container=busybox-main-container&stderr=true&stdout=true) + Feb 25 10:00:56.119: INFO: Exec stderr: "" + [AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 - Dec 14 15:58:21.519: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "job-605" for this suite. 12/14/22 15:58:21.526 + Feb 25 10:00:56.120: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-7288" for this suite. 02/25/23 10:00:56.128 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-network] Services - should complete a service status lifecycle [Conformance] - test/e2e/network/service.go:3415 -[BeforeEach] [sig-network] Services +[sig-api-machinery] Watchers + should be able to restart watching from the last resource version observed by the previous watch [Conformance] + test/e2e/apimachinery/watch.go:191 +[BeforeEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:58:21.545 -Dec 14 15:58:21.545: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 15:58:21.546 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:21.571 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:21.576 -[BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 -[It] should complete a service status lifecycle [Conformance] - test/e2e/network/service.go:3415 -STEP: creating a Service 12/14/22 15:58:21.583 -STEP: watching for the Service to be added 12/14/22 15:58:21.604 -Dec 14 15:58:21.607: INFO: Found Service test-service-g5sxk in namespace services-8006 with labels: map[test-service-static:true] & ports [{http TCP 80 {0 80 } 0}] -Dec 14 15:58:21.607: INFO: Service test-service-g5sxk created -STEP: Getting /status 12/14/22 15:58:21.607 -Dec 14 15:58:21.612: INFO: Service test-service-g5sxk has LoadBalancer: {[]} -STEP: patching the ServiceStatus 12/14/22 15:58:21.612 -STEP: watching for the Service to be patched 12/14/22 15:58:21.62 -Dec 14 15:58:21.622: INFO: observed Service test-service-g5sxk in namespace services-8006 with annotations: map[] & LoadBalancer: {[]} -Dec 14 15:58:21.622: INFO: Found Service test-service-g5sxk in namespace services-8006 with annotations: map[patchedstatus:true] & LoadBalancer: {[{203.0.113.1 []}]} -Dec 14 15:58:21.622: INFO: Service test-service-g5sxk has service status patched -STEP: updating the ServiceStatus 12/14/22 15:58:21.622 -Dec 14 15:58:21.633: INFO: updatedStatus.Conditions: []v1.Condition{v1.Condition{Type:"StatusUpdate", Status:"True", ObservedGeneration:0, LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} -STEP: watching for the Service to be updated 12/14/22 15:58:21.633 -Dec 14 15:58:21.635: INFO: Observed Service test-service-g5sxk in namespace services-8006 with annotations: map[] & Conditions: {[]} -Dec 14 15:58:21.636: INFO: Observed event: &Service{ObjectMeta:{test-service-g5sxk services-8006 90cd394c-96fc-4926-90b9-a7c5f8b53457 9474 0 2022-12-14 15:58:21 +0000 UTC map[test-service-static:true] map[patchedstatus:true] [] [] [{e2e.test Update v1 2022-12-14 15:58:21 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:test-service-static":{}}},"f:spec":{"f:internalTrafficPolicy":{},"f:ports":{".":{},"k:{\"port\":80,\"protocol\":\"TCP\"}":{".":{},"f:name":{},"f:port":{},"f:protocol":{},"f:targetPort":{}}},"f:sessionAffinity":{},"f:type":{}}} } {e2e.test Update v1 2022-12-14 15:58:21 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:patchedstatus":{}}},"f:status":{"f:loadBalancer":{"f:ingress":{}}}} status}]},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:http,Protocol:TCP,Port:80,TargetPort:{0 80 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{},ClusterIP:10.233.18.155,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.233.18.155],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{LoadBalancerIngress{IP:203.0.113.1,Hostname:,Ports:[]PortStatus{},},},},Conditions:[]Condition{},},} -Dec 14 15:58:21.636: INFO: Found Service test-service-g5sxk in namespace services-8006 with annotations: map[patchedstatus:true] & Conditions: [{StatusUpdate True 0 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] -Dec 14 15:58:21.636: INFO: Service test-service-g5sxk has service status updated -STEP: patching the service 12/14/22 15:58:21.636 -STEP: watching for the Service to be patched 12/14/22 15:58:21.653 -Dec 14 15:58:21.654: INFO: observed Service test-service-g5sxk in namespace services-8006 with labels: map[test-service-static:true] -Dec 14 15:58:21.655: INFO: observed Service test-service-g5sxk in namespace services-8006 with labels: map[test-service-static:true] -Dec 14 15:58:21.655: INFO: observed Service test-service-g5sxk in namespace services-8006 with labels: map[test-service-static:true] -Dec 14 15:58:21.656: INFO: Found Service test-service-g5sxk in namespace services-8006 with labels: map[test-service:patched test-service-static:true] -Dec 14 15:58:21.656: INFO: Service test-service-g5sxk patched -STEP: deleting the service 12/14/22 15:58:21.656 -STEP: watching for the Service to be deleted 12/14/22 15:58:21.674 -Dec 14 15:58:21.676: INFO: Observed event: ADDED -Dec 14 15:58:21.677: INFO: Observed event: MODIFIED -Dec 14 15:58:21.677: INFO: Observed event: MODIFIED -Dec 14 15:58:21.677: INFO: Observed event: MODIFIED -Dec 14 15:58:21.677: INFO: Found Service test-service-g5sxk in namespace services-8006 with labels: map[test-service:patched test-service-static:true] & annotations: map[patchedstatus:true] -Dec 14 15:58:21.677: INFO: Service test-service-g5sxk deleted -[AfterEach] [sig-network] Services +STEP: Creating a kubernetes client 02/25/23 10:00:56.151 +Feb 25 10:00:56.151: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename watch 02/25/23 10:00:56.155 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:00:56.179 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:00:56.183 +[It] should be able to restart watching from the last resource version observed by the previous watch [Conformance] + test/e2e/apimachinery/watch.go:191 +STEP: creating a watch on configmaps 02/25/23 10:00:56.187 +STEP: creating a new configmap 02/25/23 10:00:56.19 +STEP: modifying the configmap once 02/25/23 10:00:56.198 +STEP: closing the watch once it receives two notifications 02/25/23 10:00:56.213 +Feb 25 10:00:56.213: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-9114 b6c3eb38-2428-4318-8c6a-be14c1073308 13113 0 2023-02-25 10:00:56 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2023-02-25 10:00:56 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} +Feb 25 10:00:56.214: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-9114 b6c3eb38-2428-4318-8c6a-be14c1073308 13114 0 2023-02-25 10:00:56 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2023-02-25 10:00:56 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} +STEP: modifying the configmap a second time, while the watch is closed 02/25/23 10:00:56.214 +STEP: creating a new watch on configmaps from the last resource version observed by the first watch 02/25/23 10:00:56.225 +STEP: deleting the configmap 02/25/23 10:00:56.228 +STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed 02/25/23 10:00:56.237 +Feb 25 10:00:56.238: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-9114 b6c3eb38-2428-4318-8c6a-be14c1073308 13115 0 2023-02-25 10:00:56 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2023-02-25 10:00:56 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} +Feb 25 10:00:56.238: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-9114 b6c3eb38-2428-4318-8c6a-be14c1073308 13116 0 2023-02-25 10:00:56 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2023-02-25 10:00:56 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} +[AfterEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:187 -Dec 14 15:58:21.677: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-8006" for this suite. 12/14/22 15:58:21.682 -[AfterEach] [sig-network] Services - test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should complete a service status lifecycle [Conformance]","completed":53,"skipped":992,"failed":0} +Feb 25 10:00:56.238: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "watch-9114" for this suite. 02/25/23 10:00:56.245 +{"msg":"PASSED [sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance]","completed":54,"skipped":1146,"failed":0} ------------------------------ -• [0.145 seconds] -[sig-network] Services -test/e2e/network/common/framework.go:23 - should complete a service status lifecycle [Conformance] - test/e2e/network/service.go:3415 +• [0.109 seconds] +[sig-api-machinery] Watchers +test/e2e/apimachinery/framework.go:23 + should be able to restart watching from the last resource version observed by the previous watch [Conformance] + test/e2e/apimachinery/watch.go:191 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Services + [BeforeEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:58:21.545 - Dec 14 15:58:21.545: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 15:58:21.546 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:21.571 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:21.576 - [BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 - [It] should complete a service status lifecycle [Conformance] - test/e2e/network/service.go:3415 - STEP: creating a Service 12/14/22 15:58:21.583 - STEP: watching for the Service to be added 12/14/22 15:58:21.604 - Dec 14 15:58:21.607: INFO: Found Service test-service-g5sxk in namespace services-8006 with labels: map[test-service-static:true] & ports [{http TCP 80 {0 80 } 0}] - Dec 14 15:58:21.607: INFO: Service test-service-g5sxk created - STEP: Getting /status 12/14/22 15:58:21.607 - Dec 14 15:58:21.612: INFO: Service test-service-g5sxk has LoadBalancer: {[]} - STEP: patching the ServiceStatus 12/14/22 15:58:21.612 - STEP: watching for the Service to be patched 12/14/22 15:58:21.62 - Dec 14 15:58:21.622: INFO: observed Service test-service-g5sxk in namespace services-8006 with annotations: map[] & LoadBalancer: {[]} - Dec 14 15:58:21.622: INFO: Found Service test-service-g5sxk in namespace services-8006 with annotations: map[patchedstatus:true] & LoadBalancer: {[{203.0.113.1 []}]} - Dec 14 15:58:21.622: INFO: Service test-service-g5sxk has service status patched - STEP: updating the ServiceStatus 12/14/22 15:58:21.622 - Dec 14 15:58:21.633: INFO: updatedStatus.Conditions: []v1.Condition{v1.Condition{Type:"StatusUpdate", Status:"True", ObservedGeneration:0, LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} - STEP: watching for the Service to be updated 12/14/22 15:58:21.633 - Dec 14 15:58:21.635: INFO: Observed Service test-service-g5sxk in namespace services-8006 with annotations: map[] & Conditions: {[]} - Dec 14 15:58:21.636: INFO: Observed event: &Service{ObjectMeta:{test-service-g5sxk services-8006 90cd394c-96fc-4926-90b9-a7c5f8b53457 9474 0 2022-12-14 15:58:21 +0000 UTC map[test-service-static:true] map[patchedstatus:true] [] [] [{e2e.test Update v1 2022-12-14 15:58:21 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:test-service-static":{}}},"f:spec":{"f:internalTrafficPolicy":{},"f:ports":{".":{},"k:{\"port\":80,\"protocol\":\"TCP\"}":{".":{},"f:name":{},"f:port":{},"f:protocol":{},"f:targetPort":{}}},"f:sessionAffinity":{},"f:type":{}}} } {e2e.test Update v1 2022-12-14 15:58:21 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:patchedstatus":{}}},"f:status":{"f:loadBalancer":{"f:ingress":{}}}} status}]},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:http,Protocol:TCP,Port:80,TargetPort:{0 80 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{},ClusterIP:10.233.18.155,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.233.18.155],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{LoadBalancerIngress{IP:203.0.113.1,Hostname:,Ports:[]PortStatus{},},},},Conditions:[]Condition{},},} - Dec 14 15:58:21.636: INFO: Found Service test-service-g5sxk in namespace services-8006 with annotations: map[patchedstatus:true] & Conditions: [{StatusUpdate True 0 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] - Dec 14 15:58:21.636: INFO: Service test-service-g5sxk has service status updated - STEP: patching the service 12/14/22 15:58:21.636 - STEP: watching for the Service to be patched 12/14/22 15:58:21.653 - Dec 14 15:58:21.654: INFO: observed Service test-service-g5sxk in namespace services-8006 with labels: map[test-service-static:true] - Dec 14 15:58:21.655: INFO: observed Service test-service-g5sxk in namespace services-8006 with labels: map[test-service-static:true] - Dec 14 15:58:21.655: INFO: observed Service test-service-g5sxk in namespace services-8006 with labels: map[test-service-static:true] - Dec 14 15:58:21.656: INFO: Found Service test-service-g5sxk in namespace services-8006 with labels: map[test-service:patched test-service-static:true] - Dec 14 15:58:21.656: INFO: Service test-service-g5sxk patched - STEP: deleting the service 12/14/22 15:58:21.656 - STEP: watching for the Service to be deleted 12/14/22 15:58:21.674 - Dec 14 15:58:21.676: INFO: Observed event: ADDED - Dec 14 15:58:21.677: INFO: Observed event: MODIFIED - Dec 14 15:58:21.677: INFO: Observed event: MODIFIED - Dec 14 15:58:21.677: INFO: Observed event: MODIFIED - Dec 14 15:58:21.677: INFO: Found Service test-service-g5sxk in namespace services-8006 with labels: map[test-service:patched test-service-static:true] & annotations: map[patchedstatus:true] - Dec 14 15:58:21.677: INFO: Service test-service-g5sxk deleted - [AfterEach] [sig-network] Services + STEP: Creating a kubernetes client 02/25/23 10:00:56.151 + Feb 25 10:00:56.151: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename watch 02/25/23 10:00:56.155 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:00:56.179 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:00:56.183 + [It] should be able to restart watching from the last resource version observed by the previous watch [Conformance] + test/e2e/apimachinery/watch.go:191 + STEP: creating a watch on configmaps 02/25/23 10:00:56.187 + STEP: creating a new configmap 02/25/23 10:00:56.19 + STEP: modifying the configmap once 02/25/23 10:00:56.198 + STEP: closing the watch once it receives two notifications 02/25/23 10:00:56.213 + Feb 25 10:00:56.213: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-9114 b6c3eb38-2428-4318-8c6a-be14c1073308 13113 0 2023-02-25 10:00:56 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2023-02-25 10:00:56 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} + Feb 25 10:00:56.214: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-9114 b6c3eb38-2428-4318-8c6a-be14c1073308 13114 0 2023-02-25 10:00:56 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2023-02-25 10:00:56 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} + STEP: modifying the configmap a second time, while the watch is closed 02/25/23 10:00:56.214 + STEP: creating a new watch on configmaps from the last resource version observed by the first watch 02/25/23 10:00:56.225 + STEP: deleting the configmap 02/25/23 10:00:56.228 + STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed 02/25/23 10:00:56.237 + Feb 25 10:00:56.238: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-9114 b6c3eb38-2428-4318-8c6a-be14c1073308 13115 0 2023-02-25 10:00:56 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2023-02-25 10:00:56 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} + Feb 25 10:00:56.238: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-9114 b6c3eb38-2428-4318-8c6a-be14c1073308 13116 0 2023-02-25 10:00:56 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2023-02-25 10:00:56 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} + [AfterEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:187 - Dec 14 15:58:21.677: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-8006" for this suite. 12/14/22 15:58:21.682 - [AfterEach] [sig-network] Services - test/e2e/network/service.go:762 + Feb 25 10:00:56.238: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "watch-9114" for this suite. 02/25/23 10:00:56.245 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSS ------------------------------- -[sig-instrumentation] Events - should delete a collection of events [Conformance] - test/e2e/instrumentation/core_events.go:175 -[BeforeEach] [sig-instrumentation] Events +[sig-cli] Kubectl client Kubectl diff + should check if kubectl diff finds a difference for Deployments [Conformance] + test/e2e/kubectl/kubectl.go:929 +[BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:58:21.693 -Dec 14 15:58:21.694: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename events 12/14/22 15:58:21.696 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:21.718 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:21.722 -[It] should delete a collection of events [Conformance] - test/e2e/instrumentation/core_events.go:175 -STEP: Create set of events 12/14/22 15:58:21.724 -Dec 14 15:58:21.731: INFO: created test-event-1 -Dec 14 15:58:21.736: INFO: created test-event-2 -Dec 14 15:58:21.741: INFO: created test-event-3 -STEP: get a list of Events with a label in the current namespace 12/14/22 15:58:21.741 -STEP: delete collection of events 12/14/22 15:58:21.746 -Dec 14 15:58:21.746: INFO: requesting DeleteCollection of events -STEP: check that the list of events matches the requested quantity 12/14/22 15:58:21.829 -Dec 14 15:58:21.829: INFO: requesting list of events to confirm quantity -[AfterEach] [sig-instrumentation] Events +STEP: Creating a kubernetes client 02/25/23 10:00:56.263 +Feb 25 10:00:56.263: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 10:00:56.265 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:00:56.295 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:00:56.301 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 +[It] should check if kubectl diff finds a difference for Deployments [Conformance] + test/e2e/kubectl/kubectl.go:929 +STEP: create deployment with httpd image 02/25/23 10:00:56.305 +Feb 25 10:00:56.306: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1683 create -f -' +Feb 25 10:00:57.807: INFO: stderr: "" +Feb 25 10:00:57.807: INFO: stdout: "deployment.apps/httpd-deployment created\n" +STEP: verify diff finds difference between live and declared image 02/25/23 10:00:57.807 +Feb 25 10:00:57.808: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1683 diff -f -' +Feb 25 10:00:58.347: INFO: rc: 1 +Feb 25 10:00:58.348: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1683 delete -f -' +Feb 25 10:00:58.501: INFO: stderr: "" +Feb 25 10:00:58.501: INFO: stdout: "deployment.apps \"httpd-deployment\" deleted\n" +[AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 -Dec 14 15:58:21.833: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "events-4182" for this suite. 12/14/22 15:58:21.839 -{"msg":"PASSED [sig-instrumentation] Events should delete a collection of events [Conformance]","completed":54,"skipped":1000,"failed":0} +Feb 25 10:00:58.501: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-1683" for this suite. 02/25/23 10:00:58.515 +{"msg":"PASSED [sig-cli] Kubectl client Kubectl diff should check if kubectl diff finds a difference for Deployments [Conformance]","completed":55,"skipped":1146,"failed":0} ------------------------------ -• [0.154 seconds] -[sig-instrumentation] Events -test/e2e/instrumentation/common/framework.go:23 - should delete a collection of events [Conformance] - test/e2e/instrumentation/core_events.go:175 +• [2.267 seconds] +[sig-cli] Kubectl client +test/e2e/kubectl/framework.go:23 + Kubectl diff + test/e2e/kubectl/kubectl.go:923 + should check if kubectl diff finds a difference for Deployments [Conformance] + test/e2e/kubectl/kubectl.go:929 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-instrumentation] Events + [BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:58:21.693 - Dec 14 15:58:21.694: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename events 12/14/22 15:58:21.696 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:21.718 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:21.722 - [It] should delete a collection of events [Conformance] - test/e2e/instrumentation/core_events.go:175 - STEP: Create set of events 12/14/22 15:58:21.724 - Dec 14 15:58:21.731: INFO: created test-event-1 - Dec 14 15:58:21.736: INFO: created test-event-2 - Dec 14 15:58:21.741: INFO: created test-event-3 - STEP: get a list of Events with a label in the current namespace 12/14/22 15:58:21.741 - STEP: delete collection of events 12/14/22 15:58:21.746 - Dec 14 15:58:21.746: INFO: requesting DeleteCollection of events - STEP: check that the list of events matches the requested quantity 12/14/22 15:58:21.829 - Dec 14 15:58:21.829: INFO: requesting list of events to confirm quantity - [AfterEach] [sig-instrumentation] Events + STEP: Creating a kubernetes client 02/25/23 10:00:56.263 + Feb 25 10:00:56.263: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 10:00:56.265 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:00:56.295 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:00:56.301 + [BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 + [It] should check if kubectl diff finds a difference for Deployments [Conformance] + test/e2e/kubectl/kubectl.go:929 + STEP: create deployment with httpd image 02/25/23 10:00:56.305 + Feb 25 10:00:56.306: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1683 create -f -' + Feb 25 10:00:57.807: INFO: stderr: "" + Feb 25 10:00:57.807: INFO: stdout: "deployment.apps/httpd-deployment created\n" + STEP: verify diff finds difference between live and declared image 02/25/23 10:00:57.807 + Feb 25 10:00:57.808: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1683 diff -f -' + Feb 25 10:00:58.347: INFO: rc: 1 + Feb 25 10:00:58.348: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1683 delete -f -' + Feb 25 10:00:58.501: INFO: stderr: "" + Feb 25 10:00:58.501: INFO: stdout: "deployment.apps \"httpd-deployment\" deleted\n" + [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 15:58:21.833: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "events-4182" for this suite. 12/14/22 15:58:21.839 + Feb 25 10:00:58.501: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-1683" for this suite. 02/25/23 10:00:58.515 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSS +S ------------------------------ -[sig-node] InitContainer [NodeConformance] - should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] - test/e2e/common/node/init_container.go:457 -[BeforeEach] [sig-node] InitContainer [NodeConformance] +[sig-storage] EmptyDir volumes + should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:116 +[BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:58:21.854 -Dec 14 15:58:21.854: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename init-container 12/14/22 15:58:21.855 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:21.88 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:21.889 -[BeforeEach] [sig-node] InitContainer [NodeConformance] - test/e2e/common/node/init_container.go:164 -[It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] - test/e2e/common/node/init_container.go:457 -STEP: creating the pod 12/14/22 15:58:21.892 -Dec 14 15:58:21.893: INFO: PodSpec: initContainers in spec.initContainers -[AfterEach] [sig-node] InitContainer [NodeConformance] +STEP: Creating a kubernetes client 02/25/23 10:00:58.53 +Feb 25 10:00:58.530: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir 02/25/23 10:00:58.536 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:00:58.571 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:00:58.575 +[It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:116 +STEP: Creating a pod to test emptydir 0777 on tmpfs 02/25/23 10:00:58.579 +Feb 25 10:00:58.595: INFO: Waiting up to 5m0s for pod "pod-92a31d46-40ee-43b1-a81e-5c93264e01e7" in namespace "emptydir-1066" to be "Succeeded or Failed" +Feb 25 10:00:58.602: INFO: Pod "pod-92a31d46-40ee-43b1-a81e-5c93264e01e7": Phase="Pending", Reason="", readiness=false. Elapsed: 7.231949ms +Feb 25 10:01:00.611: INFO: Pod "pod-92a31d46-40ee-43b1-a81e-5c93264e01e7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016068106s +Feb 25 10:01:02.610: INFO: Pod "pod-92a31d46-40ee-43b1-a81e-5c93264e01e7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014773662s +STEP: Saw pod success 02/25/23 10:01:02.61 +Feb 25 10:01:02.610: INFO: Pod "pod-92a31d46-40ee-43b1-a81e-5c93264e01e7" satisfied condition "Succeeded or Failed" +Feb 25 10:01:02.616: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-92a31d46-40ee-43b1-a81e-5c93264e01e7 container test-container: +STEP: delete the pod 02/25/23 10:01:02.633 +Feb 25 10:01:02.661: INFO: Waiting for pod pod-92a31d46-40ee-43b1-a81e-5c93264e01e7 to disappear +Feb 25 10:01:02.666: INFO: Pod pod-92a31d46-40ee-43b1-a81e-5c93264e01e7 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 -Dec 14 15:58:25.730: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "init-container-5036" for this suite. 12/14/22 15:58:25.738 -{"msg":"PASSED [sig-node] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]","completed":55,"skipped":1026,"failed":0} +Feb 25 10:01:02.667: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-1066" for this suite. 02/25/23 10:01:02.674 +{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","completed":56,"skipped":1147,"failed":0} ------------------------------ -• [3.893 seconds] -[sig-node] InitContainer [NodeConformance] -test/e2e/common/node/framework.go:23 - should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] - test/e2e/common/node/init_container.go:457 +• [4.155 seconds] +[sig-storage] EmptyDir volumes +test/e2e/common/storage/framework.go:23 + should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:116 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] InitContainer [NodeConformance] + [BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:58:21.854 - Dec 14 15:58:21.854: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename init-container 12/14/22 15:58:21.855 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:21.88 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:21.889 - [BeforeEach] [sig-node] InitContainer [NodeConformance] - test/e2e/common/node/init_container.go:164 - [It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] - test/e2e/common/node/init_container.go:457 - STEP: creating the pod 12/14/22 15:58:21.892 - Dec 14 15:58:21.893: INFO: PodSpec: initContainers in spec.initContainers - [AfterEach] [sig-node] InitContainer [NodeConformance] + STEP: Creating a kubernetes client 02/25/23 10:00:58.53 + Feb 25 10:00:58.530: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir 02/25/23 10:00:58.536 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:00:58.571 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:00:58.575 + [It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:116 + STEP: Creating a pod to test emptydir 0777 on tmpfs 02/25/23 10:00:58.579 + Feb 25 10:00:58.595: INFO: Waiting up to 5m0s for pod "pod-92a31d46-40ee-43b1-a81e-5c93264e01e7" in namespace "emptydir-1066" to be "Succeeded or Failed" + Feb 25 10:00:58.602: INFO: Pod "pod-92a31d46-40ee-43b1-a81e-5c93264e01e7": Phase="Pending", Reason="", readiness=false. Elapsed: 7.231949ms + Feb 25 10:01:00.611: INFO: Pod "pod-92a31d46-40ee-43b1-a81e-5c93264e01e7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016068106s + Feb 25 10:01:02.610: INFO: Pod "pod-92a31d46-40ee-43b1-a81e-5c93264e01e7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014773662s + STEP: Saw pod success 02/25/23 10:01:02.61 + Feb 25 10:01:02.610: INFO: Pod "pod-92a31d46-40ee-43b1-a81e-5c93264e01e7" satisfied condition "Succeeded or Failed" + Feb 25 10:01:02.616: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-92a31d46-40ee-43b1-a81e-5c93264e01e7 container test-container: + STEP: delete the pod 02/25/23 10:01:02.633 + Feb 25 10:01:02.661: INFO: Waiting for pod pod-92a31d46-40ee-43b1-a81e-5c93264e01e7 to disappear + Feb 25 10:01:02.666: INFO: Pod pod-92a31d46-40ee-43b1-a81e-5c93264e01e7 no longer exists + [AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 - Dec 14 15:58:25.730: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "init-container-5036" for this suite. 12/14/22 15:58:25.738 + Feb 25 10:01:02.667: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-1066" for this suite. 02/25/23 10:01:02.674 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSS ------------------------------ -[sig-api-machinery] Garbage collector - should not be blocked by dependency circle [Conformance] - test/e2e/apimachinery/garbage_collector.go:849 -[BeforeEach] [sig-api-machinery] Garbage collector +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + works for CRD preserving unknown fields at the schema root [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:193 +[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:58:25.781 -Dec 14 15:58:25.782: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename gc 12/14/22 15:58:25.784 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:25.803 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:25.807 -[It] should not be blocked by dependency circle [Conformance] - test/e2e/apimachinery/garbage_collector.go:849 -Dec 14 15:58:25.849: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"da93f9a7-b09e-4e79-8601-38e2f753b808", Controller:(*bool)(0xc00344f256), BlockOwnerDeletion:(*bool)(0xc00344f257)}} -Dec 14 15:58:25.870: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"4be72f7b-785a-4b08-bc9d-ea87c3746d3b", Controller:(*bool)(0xc00344f496), BlockOwnerDeletion:(*bool)(0xc00344f497)}} -Dec 14 15:58:25.881: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"b8cdce19-5832-4701-ae6f-9eafd837826e", Controller:(*bool)(0xc00314f45e), BlockOwnerDeletion:(*bool)(0xc00314f45f)}} -[AfterEach] [sig-api-machinery] Garbage collector +STEP: Creating a kubernetes client 02/25/23 10:01:02.686 +Feb 25 10:01:02.687: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 10:01:02.689 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:02.713 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:02.717 +[It] works for CRD preserving unknown fields at the schema root [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:193 +Feb 25 10:01:02.721: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: kubectl validation (kubectl create and apply) allows request with any unknown properties 02/25/23 10:01:08.831 +Feb 25 10:01:08.832: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-4025 --namespace=crd-publish-openapi-4025 create -f -' +Feb 25 10:01:10.566: INFO: stderr: "" +Feb 25 10:01:10.566: INFO: stdout: "e2e-test-crd-publish-openapi-6808-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n" +Feb 25 10:01:10.567: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-4025 --namespace=crd-publish-openapi-4025 delete e2e-test-crd-publish-openapi-6808-crds test-cr' +Feb 25 10:01:10.724: INFO: stderr: "" +Feb 25 10:01:10.724: INFO: stdout: "e2e-test-crd-publish-openapi-6808-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n" +Feb 25 10:01:10.724: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-4025 --namespace=crd-publish-openapi-4025 apply -f -' +Feb 25 10:01:12.010: INFO: stderr: "" +Feb 25 10:01:12.010: INFO: stdout: "e2e-test-crd-publish-openapi-6808-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n" +Feb 25 10:01:12.011: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-4025 --namespace=crd-publish-openapi-4025 delete e2e-test-crd-publish-openapi-6808-crds test-cr' +Feb 25 10:01:12.194: INFO: stderr: "" +Feb 25 10:01:12.194: INFO: stdout: "e2e-test-crd-publish-openapi-6808-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n" +STEP: kubectl explain works to explain CR 02/25/23 10:01:12.194 +Feb 25 10:01:12.195: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-4025 explain e2e-test-crd-publish-openapi-6808-crds' +Feb 25 10:01:12.678: INFO: stderr: "" +Feb 25 10:01:12.678: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-6808-crd\nVERSION: crd-publish-openapi-test-unknown-at-root.example.com/v1\n\nDESCRIPTION:\n \n" +[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 15:58:30.902: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "gc-6511" for this suite. 12/14/22 15:58:30.917 -{"msg":"PASSED [sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance]","completed":56,"skipped":1085,"failed":0} +Feb 25 10:01:18.659: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "crd-publish-openapi-4025" for this suite. 02/25/23 10:01:18.676 +{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields at the schema root [Conformance]","completed":57,"skipped":1154,"failed":0} ------------------------------ -• [SLOW TEST] [5.145 seconds] -[sig-api-machinery] Garbage collector +• [SLOW TEST] [16.002 seconds] +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/apimachinery/framework.go:23 - should not be blocked by dependency circle [Conformance] - test/e2e/apimachinery/garbage_collector.go:849 + works for CRD preserving unknown fields at the schema root [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:193 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Garbage collector + [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:58:25.781 - Dec 14 15:58:25.782: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename gc 12/14/22 15:58:25.784 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:25.803 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:25.807 - [It] should not be blocked by dependency circle [Conformance] - test/e2e/apimachinery/garbage_collector.go:849 - Dec 14 15:58:25.849: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"da93f9a7-b09e-4e79-8601-38e2f753b808", Controller:(*bool)(0xc00344f256), BlockOwnerDeletion:(*bool)(0xc00344f257)}} - Dec 14 15:58:25.870: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"4be72f7b-785a-4b08-bc9d-ea87c3746d3b", Controller:(*bool)(0xc00344f496), BlockOwnerDeletion:(*bool)(0xc00344f497)}} - Dec 14 15:58:25.881: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"b8cdce19-5832-4701-ae6f-9eafd837826e", Controller:(*bool)(0xc00314f45e), BlockOwnerDeletion:(*bool)(0xc00314f45f)}} - [AfterEach] [sig-api-machinery] Garbage collector + STEP: Creating a kubernetes client 02/25/23 10:01:02.686 + Feb 25 10:01:02.687: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 10:01:02.689 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:02.713 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:02.717 + [It] works for CRD preserving unknown fields at the schema root [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:193 + Feb 25 10:01:02.721: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: kubectl validation (kubectl create and apply) allows request with any unknown properties 02/25/23 10:01:08.831 + Feb 25 10:01:08.832: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-4025 --namespace=crd-publish-openapi-4025 create -f -' + Feb 25 10:01:10.566: INFO: stderr: "" + Feb 25 10:01:10.566: INFO: stdout: "e2e-test-crd-publish-openapi-6808-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n" + Feb 25 10:01:10.567: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-4025 --namespace=crd-publish-openapi-4025 delete e2e-test-crd-publish-openapi-6808-crds test-cr' + Feb 25 10:01:10.724: INFO: stderr: "" + Feb 25 10:01:10.724: INFO: stdout: "e2e-test-crd-publish-openapi-6808-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n" + Feb 25 10:01:10.724: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-4025 --namespace=crd-publish-openapi-4025 apply -f -' + Feb 25 10:01:12.010: INFO: stderr: "" + Feb 25 10:01:12.010: INFO: stdout: "e2e-test-crd-publish-openapi-6808-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n" + Feb 25 10:01:12.011: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-4025 --namespace=crd-publish-openapi-4025 delete e2e-test-crd-publish-openapi-6808-crds test-cr' + Feb 25 10:01:12.194: INFO: stderr: "" + Feb 25 10:01:12.194: INFO: stdout: "e2e-test-crd-publish-openapi-6808-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n" + STEP: kubectl explain works to explain CR 02/25/23 10:01:12.194 + Feb 25 10:01:12.195: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-4025 explain e2e-test-crd-publish-openapi-6808-crds' + Feb 25 10:01:12.678: INFO: stderr: "" + Feb 25 10:01:12.678: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-6808-crd\nVERSION: crd-publish-openapi-test-unknown-at-root.example.com/v1\n\nDESCRIPTION:\n \n" + [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 15:58:30.902: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "gc-6511" for this suite. 12/14/22 15:58:30.917 + Feb 25 10:01:18.659: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "crd-publish-openapi-4025" for this suite. 02/25/23 10:01:18.676 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSS +SSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Containers - should be able to override the image's default arguments (container cmd) [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:58 -[BeforeEach] [sig-node] Containers +[sig-node] Downward API + should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:43 +[BeforeEach] [sig-node] Downward API test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:58:30.931 -Dec 14 15:58:30.931: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename containers 12/14/22 15:58:30.935 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:30.958 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:30.962 -[It] should be able to override the image's default arguments (container cmd) [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:58 -STEP: Creating a pod to test override arguments 12/14/22 15:58:30.967 -Dec 14 15:58:30.991: INFO: Waiting up to 5m0s for pod "client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014" in namespace "containers-1651" to be "Succeeded or Failed" -Dec 14 15:58:30.997: INFO: Pod "client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014": Phase="Pending", Reason="", readiness=false. Elapsed: 6.275588ms -Dec 14 15:58:33.004: INFO: Pod "client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013316926s -Dec 14 15:58:35.006: INFO: Pod "client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015738784s -STEP: Saw pod success 12/14/22 15:58:35.006 -Dec 14 15:58:35.007: INFO: Pod "client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014" satisfied condition "Succeeded or Failed" -Dec 14 15:58:35.013: INFO: Trying to get logs from node kiedai4iakae-3 pod client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014 container agnhost-container: -STEP: delete the pod 12/14/22 15:58:35.025 -Dec 14 15:58:35.045: INFO: Waiting for pod client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014 to disappear -Dec 14 15:58:35.049: INFO: Pod client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014 no longer exists -[AfterEach] [sig-node] Containers +STEP: Creating a kubernetes client 02/25/23 10:01:18.698 +Feb 25 10:01:18.698: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 10:01:18.699 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:18.736 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:18.743 +[It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:43 +STEP: Creating a pod to test downward api env vars 02/25/23 10:01:18.749 +Feb 25 10:01:18.779: INFO: Waiting up to 5m0s for pod "downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a" in namespace "downward-api-636" to be "Succeeded or Failed" +Feb 25 10:01:18.792: INFO: Pod "downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.352681ms +Feb 25 10:01:20.800: INFO: Pod "downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021338311s +Feb 25 10:01:22.801: INFO: Pod "downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022442908s +STEP: Saw pod success 02/25/23 10:01:22.802 +Feb 25 10:01:22.802: INFO: Pod "downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a" satisfied condition "Succeeded or Failed" +Feb 25 10:01:22.809: INFO: Trying to get logs from node diek3dua9gay-3 pod downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a container dapi-container: +STEP: delete the pod 02/25/23 10:01:22.839 +Feb 25 10:01:22.861: INFO: Waiting for pod downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a to disappear +Feb 25 10:01:22.867: INFO: Pod downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a no longer exists +[AfterEach] [sig-node] Downward API test/e2e/framework/framework.go:187 -Dec 14 15:58:35.050: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "containers-1651" for this suite. 12/14/22 15:58:35.055 -{"msg":"PASSED [sig-node] Containers should be able to override the image's default arguments (container cmd) [NodeConformance] [Conformance]","completed":57,"skipped":1092,"failed":0} +Feb 25 10:01:22.867: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-636" for this suite. 02/25/23 10:01:22.886 +{"msg":"PASSED [sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]","completed":58,"skipped":1174,"failed":0} ------------------------------ -• [4.146 seconds] -[sig-node] Containers +• [4.220 seconds] +[sig-node] Downward API test/e2e/common/node/framework.go:23 - should be able to override the image's default arguments (container cmd) [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:58 - - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Containers - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:58:30.931 - Dec 14 15:58:30.931: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename containers 12/14/22 15:58:30.935 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:30.958 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:30.962 - [It] should be able to override the image's default arguments (container cmd) [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:58 - STEP: Creating a pod to test override arguments 12/14/22 15:58:30.967 - Dec 14 15:58:30.991: INFO: Waiting up to 5m0s for pod "client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014" in namespace "containers-1651" to be "Succeeded or Failed" - Dec 14 15:58:30.997: INFO: Pod "client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014": Phase="Pending", Reason="", readiness=false. Elapsed: 6.275588ms - Dec 14 15:58:33.004: INFO: Pod "client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013316926s - Dec 14 15:58:35.006: INFO: Pod "client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015738784s - STEP: Saw pod success 12/14/22 15:58:35.006 - Dec 14 15:58:35.007: INFO: Pod "client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014" satisfied condition "Succeeded or Failed" - Dec 14 15:58:35.013: INFO: Trying to get logs from node kiedai4iakae-3 pod client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014 container agnhost-container: - STEP: delete the pod 12/14/22 15:58:35.025 - Dec 14 15:58:35.045: INFO: Waiting for pod client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014 to disappear - Dec 14 15:58:35.049: INFO: Pod client-containers-097357cf-49f8-4d9b-94ef-2c4ed7c6c014 no longer exists - [AfterEach] [sig-node] Containers - test/e2e/framework/framework.go:187 - Dec 14 15:58:35.050: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "containers-1651" for this suite. 12/14/22 15:58:35.055 - << End Captured GinkgoWriter Output ------------------------------- -SSS ------------------------------- -[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] - should validate Statefulset Status endpoints [Conformance] - test/e2e/apps/statefulset.go:975 -[BeforeEach] [sig-apps] StatefulSet - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:58:35.079 -Dec 14 15:58:35.079: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename statefulset 12/14/22 15:58:35.083 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:35.129 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:35.136 -[BeforeEach] [sig-apps] StatefulSet - test/e2e/apps/statefulset.go:96 -[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:111 -STEP: Creating service test in namespace statefulset-5042 12/14/22 15:58:35.141 -[It] should validate Statefulset Status endpoints [Conformance] - test/e2e/apps/statefulset.go:975 -STEP: Creating statefulset ss in namespace statefulset-5042 12/14/22 15:58:35.17 -Dec 14 15:58:35.195: INFO: Found 0 stateful pods, waiting for 1 -Dec 14 15:58:45.202: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true -STEP: Patch Statefulset to include a label 12/14/22 15:58:45.211 -STEP: Getting /status 12/14/22 15:58:45.228 -Dec 14 15:58:45.234: INFO: StatefulSet ss has Conditions: []v1.StatefulSetCondition(nil) -STEP: updating the StatefulSet Status 12/14/22 15:58:45.235 -Dec 14 15:58:45.249: INFO: updatedStatus.Conditions: []v1.StatefulSetCondition{v1.StatefulSetCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} -STEP: watching for the statefulset status to be updated 12/14/22 15:58:45.249 -Dec 14 15:58:45.253: INFO: Observed &StatefulSet event: ADDED -Dec 14 15:58:45.253: INFO: Found Statefulset ss in namespace statefulset-5042 with labels: map[e2e:testing] annotations: map[] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} -Dec 14 15:58:45.253: INFO: Statefulset ss has an updated status -STEP: patching the Statefulset Status 12/14/22 15:58:45.253 -Dec 14 15:58:45.254: INFO: Patch payload: {"status":{"conditions":[{"type":"StatusPatched","status":"True"}]}} -Dec 14 15:58:45.261: INFO: Patched status conditions: []v1.StatefulSetCondition{v1.StatefulSetCondition{Type:"StatusPatched", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"", Message:""}} -STEP: watching for the Statefulset status to be patched 12/14/22 15:58:45.261 -Dec 14 15:58:45.267: INFO: Observed &StatefulSet event: ADDED -[AfterEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:122 -Dec 14 15:58:45.268: INFO: Deleting all statefulset in ns statefulset-5042 -Dec 14 15:58:45.272: INFO: Scaling statefulset ss to 0 -Dec 14 15:58:55.297: INFO: Waiting for statefulset status.replicas updated to 0 -Dec 14 15:58:55.303: INFO: Deleting statefulset ss -[AfterEach] [sig-apps] StatefulSet - test/e2e/framework/framework.go:187 -Dec 14 15:58:55.329: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "statefulset-5042" for this suite. 12/14/22 15:58:55.339 -{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should validate Statefulset Status endpoints [Conformance]","completed":58,"skipped":1095,"failed":0} ------------------------------- -• [SLOW TEST] [20.269 seconds] -[sig-apps] StatefulSet -test/e2e/apps/framework.go:23 - Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:101 - should validate Statefulset Status endpoints [Conformance] - test/e2e/apps/statefulset.go:975 + should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:43 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] StatefulSet + [BeforeEach] [sig-node] Downward API test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:58:35.079 - Dec 14 15:58:35.079: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename statefulset 12/14/22 15:58:35.083 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:35.129 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:35.136 - [BeforeEach] [sig-apps] StatefulSet - test/e2e/apps/statefulset.go:96 - [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:111 - STEP: Creating service test in namespace statefulset-5042 12/14/22 15:58:35.141 - [It] should validate Statefulset Status endpoints [Conformance] - test/e2e/apps/statefulset.go:975 - STEP: Creating statefulset ss in namespace statefulset-5042 12/14/22 15:58:35.17 - Dec 14 15:58:35.195: INFO: Found 0 stateful pods, waiting for 1 - Dec 14 15:58:45.202: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true - STEP: Patch Statefulset to include a label 12/14/22 15:58:45.211 - STEP: Getting /status 12/14/22 15:58:45.228 - Dec 14 15:58:45.234: INFO: StatefulSet ss has Conditions: []v1.StatefulSetCondition(nil) - STEP: updating the StatefulSet Status 12/14/22 15:58:45.235 - Dec 14 15:58:45.249: INFO: updatedStatus.Conditions: []v1.StatefulSetCondition{v1.StatefulSetCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} - STEP: watching for the statefulset status to be updated 12/14/22 15:58:45.249 - Dec 14 15:58:45.253: INFO: Observed &StatefulSet event: ADDED - Dec 14 15:58:45.253: INFO: Found Statefulset ss in namespace statefulset-5042 with labels: map[e2e:testing] annotations: map[] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} - Dec 14 15:58:45.253: INFO: Statefulset ss has an updated status - STEP: patching the Statefulset Status 12/14/22 15:58:45.253 - Dec 14 15:58:45.254: INFO: Patch payload: {"status":{"conditions":[{"type":"StatusPatched","status":"True"}]}} - Dec 14 15:58:45.261: INFO: Patched status conditions: []v1.StatefulSetCondition{v1.StatefulSetCondition{Type:"StatusPatched", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"", Message:""}} - STEP: watching for the Statefulset status to be patched 12/14/22 15:58:45.261 - Dec 14 15:58:45.267: INFO: Observed &StatefulSet event: ADDED - [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:122 - Dec 14 15:58:45.268: INFO: Deleting all statefulset in ns statefulset-5042 - Dec 14 15:58:45.272: INFO: Scaling statefulset ss to 0 - Dec 14 15:58:55.297: INFO: Waiting for statefulset status.replicas updated to 0 - Dec 14 15:58:55.303: INFO: Deleting statefulset ss - [AfterEach] [sig-apps] StatefulSet + STEP: Creating a kubernetes client 02/25/23 10:01:18.698 + Feb 25 10:01:18.698: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 10:01:18.699 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:18.736 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:18.743 + [It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:43 + STEP: Creating a pod to test downward api env vars 02/25/23 10:01:18.749 + Feb 25 10:01:18.779: INFO: Waiting up to 5m0s for pod "downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a" in namespace "downward-api-636" to be "Succeeded or Failed" + Feb 25 10:01:18.792: INFO: Pod "downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.352681ms + Feb 25 10:01:20.800: INFO: Pod "downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021338311s + Feb 25 10:01:22.801: INFO: Pod "downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022442908s + STEP: Saw pod success 02/25/23 10:01:22.802 + Feb 25 10:01:22.802: INFO: Pod "downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a" satisfied condition "Succeeded or Failed" + Feb 25 10:01:22.809: INFO: Trying to get logs from node diek3dua9gay-3 pod downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a container dapi-container: + STEP: delete the pod 02/25/23 10:01:22.839 + Feb 25 10:01:22.861: INFO: Waiting for pod downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a to disappear + Feb 25 10:01:22.867: INFO: Pod downward-api-aa97e2c3-3c80-46d7-81c2-7186c47da23a no longer exists + [AfterEach] [sig-node] Downward API test/e2e/framework/framework.go:187 - Dec 14 15:58:55.329: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "statefulset-5042" for this suite. 12/14/22 15:58:55.339 + Feb 25 10:01:22.867: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-636" for this suite. 02/25/23 10:01:22.886 << End Captured GinkgoWriter Output ------------------------------ -SSSS +SSSSSSSSSS ------------------------------ -[sig-node] Kubelet when scheduling a busybox command in a pod - should print the output to logs [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:52 -[BeforeEach] [sig-node] Kubelet +[sig-auth] ServiceAccounts + should guarantee kube-root-ca.crt exist in any namespace [Conformance] + test/e2e/auth/service_accounts.go:739 +[BeforeEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:58:55.348 -Dec 14 15:58:55.349: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubelet-test 12/14/22 15:58:55.352 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:55.38 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:55.383 -[BeforeEach] [sig-node] Kubelet - test/e2e/common/node/kubelet.go:41 -[It] should print the output to logs [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:52 -Dec 14 15:58:55.401: INFO: Waiting up to 5m0s for pod "busybox-scheduling-819ac6f3-5de3-4e8c-b666-169628f8a3a1" in namespace "kubelet-test-2386" to be "running and ready" -Dec 14 15:58:55.406: INFO: Pod "busybox-scheduling-819ac6f3-5de3-4e8c-b666-169628f8a3a1": Phase="Pending", Reason="", readiness=false. Elapsed: 4.455695ms -Dec 14 15:58:55.406: INFO: The phase of Pod busybox-scheduling-819ac6f3-5de3-4e8c-b666-169628f8a3a1 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 15:58:57.417: INFO: Pod "busybox-scheduling-819ac6f3-5de3-4e8c-b666-169628f8a3a1": Phase="Running", Reason="", readiness=true. Elapsed: 2.015381076s -Dec 14 15:58:57.417: INFO: The phase of Pod busybox-scheduling-819ac6f3-5de3-4e8c-b666-169628f8a3a1 is Running (Ready = true) -Dec 14 15:58:57.417: INFO: Pod "busybox-scheduling-819ac6f3-5de3-4e8c-b666-169628f8a3a1" satisfied condition "running and ready" -[AfterEach] [sig-node] Kubelet +STEP: Creating a kubernetes client 02/25/23 10:01:22.927 +Feb 25 10:01:22.928: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename svcaccounts 02/25/23 10:01:22.932 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:22.968 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:22.975 +[It] should guarantee kube-root-ca.crt exist in any namespace [Conformance] + test/e2e/auth/service_accounts.go:739 +Feb 25 10:01:22.989: INFO: Got root ca configmap in namespace "svcaccounts-9528" +Feb 25 10:01:23.015: INFO: Deleted root ca configmap in namespace "svcaccounts-9528" +STEP: waiting for a new root ca configmap created 02/25/23 10:01:23.516 +Feb 25 10:01:23.523: INFO: Recreated root ca configmap in namespace "svcaccounts-9528" +Feb 25 10:01:23.533: INFO: Updated root ca configmap in namespace "svcaccounts-9528" +STEP: waiting for the root ca configmap reconciled 02/25/23 10:01:24.033 +Feb 25 10:01:24.041: INFO: Reconciled root ca configmap in namespace "svcaccounts-9528" +[AfterEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:187 -Dec 14 15:58:57.437: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubelet-test-2386" for this suite. 12/14/22 15:58:57.445 -{"msg":"PASSED [sig-node] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance]","completed":59,"skipped":1099,"failed":0} +Feb 25 10:01:24.042: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "svcaccounts-9528" for this suite. 02/25/23 10:01:24.053 +{"msg":"PASSED [sig-auth] ServiceAccounts should guarantee kube-root-ca.crt exist in any namespace [Conformance]","completed":59,"skipped":1184,"failed":0} ------------------------------ -• [2.119 seconds] -[sig-node] Kubelet -test/e2e/common/node/framework.go:23 - when scheduling a busybox command in a pod - test/e2e/common/node/kubelet.go:44 - should print the output to logs [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:52 +• [1.144 seconds] +[sig-auth] ServiceAccounts +test/e2e/auth/framework.go:23 + should guarantee kube-root-ca.crt exist in any namespace [Conformance] + test/e2e/auth/service_accounts.go:739 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Kubelet + [BeforeEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:58:55.348 - Dec 14 15:58:55.349: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubelet-test 12/14/22 15:58:55.352 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:55.38 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:55.383 - [BeforeEach] [sig-node] Kubelet - test/e2e/common/node/kubelet.go:41 - [It] should print the output to logs [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:52 - Dec 14 15:58:55.401: INFO: Waiting up to 5m0s for pod "busybox-scheduling-819ac6f3-5de3-4e8c-b666-169628f8a3a1" in namespace "kubelet-test-2386" to be "running and ready" - Dec 14 15:58:55.406: INFO: Pod "busybox-scheduling-819ac6f3-5de3-4e8c-b666-169628f8a3a1": Phase="Pending", Reason="", readiness=false. Elapsed: 4.455695ms - Dec 14 15:58:55.406: INFO: The phase of Pod busybox-scheduling-819ac6f3-5de3-4e8c-b666-169628f8a3a1 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 15:58:57.417: INFO: Pod "busybox-scheduling-819ac6f3-5de3-4e8c-b666-169628f8a3a1": Phase="Running", Reason="", readiness=true. Elapsed: 2.015381076s - Dec 14 15:58:57.417: INFO: The phase of Pod busybox-scheduling-819ac6f3-5de3-4e8c-b666-169628f8a3a1 is Running (Ready = true) - Dec 14 15:58:57.417: INFO: Pod "busybox-scheduling-819ac6f3-5de3-4e8c-b666-169628f8a3a1" satisfied condition "running and ready" - [AfterEach] [sig-node] Kubelet + STEP: Creating a kubernetes client 02/25/23 10:01:22.927 + Feb 25 10:01:22.928: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename svcaccounts 02/25/23 10:01:22.932 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:22.968 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:22.975 + [It] should guarantee kube-root-ca.crt exist in any namespace [Conformance] + test/e2e/auth/service_accounts.go:739 + Feb 25 10:01:22.989: INFO: Got root ca configmap in namespace "svcaccounts-9528" + Feb 25 10:01:23.015: INFO: Deleted root ca configmap in namespace "svcaccounts-9528" + STEP: waiting for a new root ca configmap created 02/25/23 10:01:23.516 + Feb 25 10:01:23.523: INFO: Recreated root ca configmap in namespace "svcaccounts-9528" + Feb 25 10:01:23.533: INFO: Updated root ca configmap in namespace "svcaccounts-9528" + STEP: waiting for the root ca configmap reconciled 02/25/23 10:01:24.033 + Feb 25 10:01:24.041: INFO: Reconciled root ca configmap in namespace "svcaccounts-9528" + [AfterEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:187 - Dec 14 15:58:57.437: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubelet-test-2386" for this suite. 12/14/22 15:58:57.445 + Feb 25 10:01:24.042: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "svcaccounts-9528" for this suite. 02/25/23 10:01:24.053 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSS +SSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - works for multiple CRDs of same group but different versions [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:308 -[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +[sig-storage] EmptyDir volumes + should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:186 +[BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:58:57.47 -Dec 14 15:58:57.470: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 15:58:57.476 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:57.496 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:57.503 -[It] works for multiple CRDs of same group but different versions [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:308 -STEP: CRs in the same group but different versions (one multiversion CRD) show up in OpenAPI documentation 12/14/22 15:58:57.509 -Dec 14 15:58:57.510: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: CRs in the same group but different versions (two CRDs) show up in OpenAPI documentation 12/14/22 15:59:13.742 -Dec 14 15:59:13.744: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 15:59:17.164: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 10:01:24.074 +Feb 25 10:01:24.075: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir 02/25/23 10:01:24.077 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:24.106 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:24.11 +[It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:186 +STEP: Creating a pod to test emptydir 0777 on node default medium 02/25/23 10:01:24.113 +Feb 25 10:01:24.129: INFO: Waiting up to 5m0s for pod "pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2" in namespace "emptydir-3703" to be "Succeeded or Failed" +Feb 25 10:01:24.136: INFO: Pod "pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2": Phase="Pending", Reason="", readiness=false. Elapsed: 6.746117ms +Feb 25 10:01:26.147: INFO: Pod "pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017488535s +Feb 25 10:01:28.143: INFO: Pod "pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013209564s +STEP: Saw pod success 02/25/23 10:01:28.143 +Feb 25 10:01:28.144: INFO: Pod "pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2" satisfied condition "Succeeded or Failed" +Feb 25 10:01:28.150: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2 container test-container: +STEP: delete the pod 02/25/23 10:01:28.163 +Feb 25 10:01:28.193: INFO: Waiting for pod pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2 to disappear +Feb 25 10:01:28.199: INFO: Pod pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 -Dec 14 15:59:31.321: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-publish-openapi-6498" for this suite. 12/14/22 15:59:31.337 -{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group but different versions [Conformance]","completed":60,"skipped":1107,"failed":0} +Feb 25 10:01:28.200: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-3703" for this suite. 02/25/23 10:01:28.209 +{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","completed":60,"skipped":1201,"failed":0} ------------------------------ -• [SLOW TEST] [33.880 seconds] -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - works for multiple CRDs of same group but different versions [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:308 +• [4.151 seconds] +[sig-storage] EmptyDir volumes +test/e2e/common/storage/framework.go:23 + should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:186 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + [BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:58:57.47 - Dec 14 15:58:57.470: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 15:58:57.476 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:58:57.496 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:58:57.503 - [It] works for multiple CRDs of same group but different versions [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:308 - STEP: CRs in the same group but different versions (one multiversion CRD) show up in OpenAPI documentation 12/14/22 15:58:57.509 - Dec 14 15:58:57.510: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: CRs in the same group but different versions (two CRDs) show up in OpenAPI documentation 12/14/22 15:59:13.742 - Dec 14 15:59:13.744: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 15:59:17.164: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 10:01:24.074 + Feb 25 10:01:24.075: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir 02/25/23 10:01:24.077 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:24.106 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:24.11 + [It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:186 + STEP: Creating a pod to test emptydir 0777 on node default medium 02/25/23 10:01:24.113 + Feb 25 10:01:24.129: INFO: Waiting up to 5m0s for pod "pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2" in namespace "emptydir-3703" to be "Succeeded or Failed" + Feb 25 10:01:24.136: INFO: Pod "pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2": Phase="Pending", Reason="", readiness=false. Elapsed: 6.746117ms + Feb 25 10:01:26.147: INFO: Pod "pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017488535s + Feb 25 10:01:28.143: INFO: Pod "pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013209564s + STEP: Saw pod success 02/25/23 10:01:28.143 + Feb 25 10:01:28.144: INFO: Pod "pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2" satisfied condition "Succeeded or Failed" + Feb 25 10:01:28.150: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2 container test-container: + STEP: delete the pod 02/25/23 10:01:28.163 + Feb 25 10:01:28.193: INFO: Waiting for pod pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2 to disappear + Feb 25 10:01:28.199: INFO: Pod pod-c2468f84-8716-4dd3-aceb-5b55e8d339d2 no longer exists + [AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 - Dec 14 15:59:31.321: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "crd-publish-openapi-6498" for this suite. 12/14/22 15:59:31.337 + Feb 25 10:01:28.200: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-3703" for this suite. 02/25/23 10:01:28.209 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSS +SSSSSSSSSSSSSSS ------------------------------ -[sig-node] Container Runtime blackbox test on terminated container - should report termination message from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:247 -[BeforeEach] [sig-node] Container Runtime +[sig-node] KubeletManagedEtcHosts + should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/kubelet_etc_hosts.go:63 +[BeforeEach] [sig-node] KubeletManagedEtcHosts test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:59:31.353 -Dec 14 15:59:31.353: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-runtime 12/14/22 15:59:31.356 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:31.389 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:31.393 -[It] should report termination message from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:247 -STEP: create the container 12/14/22 15:59:31.396 -STEP: wait for the container to reach Succeeded 12/14/22 15:59:31.409 -STEP: get the container status 12/14/22 15:59:35.459 -STEP: the container should be terminated 12/14/22 15:59:35.466 -STEP: the termination message should be set 12/14/22 15:59:35.466 -Dec 14 15:59:35.466: INFO: Expected: &{OK} to match Container's Termination Message: OK -- -STEP: delete the container 12/14/22 15:59:35.466 -[AfterEach] [sig-node] Container Runtime +STEP: Creating a kubernetes client 02/25/23 10:01:28.235 +Feb 25 10:01:28.235: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts 02/25/23 10:01:28.238 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:28.269 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:28.274 +[It] should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/kubelet_etc_hosts.go:63 +STEP: Setting up the test 02/25/23 10:01:28.279 +STEP: Creating hostNetwork=false pod 02/25/23 10:01:28.279 +Feb 25 10:01:28.295: INFO: Waiting up to 5m0s for pod "test-pod" in namespace "e2e-kubelet-etc-hosts-7469" to be "running and ready" +Feb 25 10:01:28.303: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 7.653424ms +Feb 25 10:01:28.303: INFO: The phase of Pod test-pod is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:01:30.310: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015230542s +Feb 25 10:01:30.311: INFO: The phase of Pod test-pod is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:01:32.309: INFO: Pod "test-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.014098589s +Feb 25 10:01:32.309: INFO: The phase of Pod test-pod is Running (Ready = true) +Feb 25 10:01:32.310: INFO: Pod "test-pod" satisfied condition "running and ready" +STEP: Creating hostNetwork=true pod 02/25/23 10:01:32.314 +Feb 25 10:01:32.323: INFO: Waiting up to 5m0s for pod "test-host-network-pod" in namespace "e2e-kubelet-etc-hosts-7469" to be "running and ready" +Feb 25 10:01:32.347: INFO: Pod "test-host-network-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 23.46451ms +Feb 25 10:01:32.347: INFO: The phase of Pod test-host-network-pod is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:01:34.354: INFO: Pod "test-host-network-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.031042507s +Feb 25 10:01:34.355: INFO: The phase of Pod test-host-network-pod is Running (Ready = true) +Feb 25 10:01:34.355: INFO: Pod "test-host-network-pod" satisfied condition "running and ready" +STEP: Running the test 02/25/23 10:01:34.362 +STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false 02/25/23 10:01:34.362 +Feb 25 10:01:34.363: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:01:34.363: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:01:34.367: INFO: ExecWithOptions: Clientset creation +Feb 25 10:01:34.368: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-1&container=busybox-1&stderr=true&stdout=true) +Feb 25 10:01:34.494: INFO: Exec stderr: "" +Feb 25 10:01:34.495: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:01:34.495: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:01:34.498: INFO: ExecWithOptions: Clientset creation +Feb 25 10:01:34.498: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-1&container=busybox-1&stderr=true&stdout=true) +Feb 25 10:01:34.580: INFO: Exec stderr: "" +Feb 25 10:01:34.580: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:01:34.581: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:01:34.583: INFO: ExecWithOptions: Clientset creation +Feb 25 10:01:34.583: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-2&container=busybox-2&stderr=true&stdout=true) +Feb 25 10:01:34.687: INFO: Exec stderr: "" +Feb 25 10:01:34.687: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:01:34.687: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:01:34.690: INFO: ExecWithOptions: Clientset creation +Feb 25 10:01:34.690: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-2&container=busybox-2&stderr=true&stdout=true) +Feb 25 10:01:34.770: INFO: Exec stderr: "" +STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount 02/25/23 10:01:34.77 +Feb 25 10:01:34.771: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:01:34.771: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:01:34.776: INFO: ExecWithOptions: Clientset creation +Feb 25 10:01:34.776: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-3&container=busybox-3&stderr=true&stdout=true) +Feb 25 10:01:34.917: INFO: Exec stderr: "" +Feb 25 10:01:34.917: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:01:34.917: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:01:34.920: INFO: ExecWithOptions: Clientset creation +Feb 25 10:01:34.920: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-3&container=busybox-3&stderr=true&stdout=true) +Feb 25 10:01:35.008: INFO: Exec stderr: "" +STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true 02/25/23 10:01:35.008 +Feb 25 10:01:35.008: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:01:35.008: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:01:35.010: INFO: ExecWithOptions: Clientset creation +Feb 25 10:01:35.010: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-1&container=busybox-1&stderr=true&stdout=true) +Feb 25 10:01:35.145: INFO: Exec stderr: "" +Feb 25 10:01:35.145: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:01:35.145: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:01:35.146: INFO: ExecWithOptions: Clientset creation +Feb 25 10:01:35.146: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-1&container=busybox-1&stderr=true&stdout=true) +Feb 25 10:01:35.284: INFO: Exec stderr: "" +Feb 25 10:01:35.284: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:01:35.285: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:01:35.287: INFO: ExecWithOptions: Clientset creation +Feb 25 10:01:35.287: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-2&container=busybox-2&stderr=true&stdout=true) +Feb 25 10:01:35.385: INFO: Exec stderr: "" +Feb 25 10:01:35.386: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:01:35.386: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:01:35.388: INFO: ExecWithOptions: Clientset creation +Feb 25 10:01:35.388: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-2&container=busybox-2&stderr=true&stdout=true) +Feb 25 10:01:35.477: INFO: Exec stderr: "" +[AfterEach] [sig-node] KubeletManagedEtcHosts test/e2e/framework/framework.go:187 -Dec 14 15:59:35.490: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-runtime-4074" for this suite. 12/14/22 15:59:35.495 -{"msg":"PASSED [sig-node] Container Runtime blackbox test on terminated container should report termination message from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","completed":61,"skipped":1119,"failed":0} +Feb 25 10:01:35.477: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "e2e-kubelet-etc-hosts-7469" for this suite. 02/25/23 10:01:35.486 +{"msg":"PASSED [sig-node] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]","completed":61,"skipped":1216,"failed":0} ------------------------------ -• [4.148 seconds] -[sig-node] Container Runtime +• [SLOW TEST] [7.264 seconds] +[sig-node] KubeletManagedEtcHosts test/e2e/common/node/framework.go:23 - blackbox test - test/e2e/common/node/runtime.go:43 - on terminated container - test/e2e/common/node/runtime.go:136 - should report termination message from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:247 + should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/kubelet_etc_hosts.go:63 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Container Runtime + [BeforeEach] [sig-node] KubeletManagedEtcHosts test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:59:31.353 - Dec 14 15:59:31.353: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-runtime 12/14/22 15:59:31.356 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:31.389 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:31.393 - [It] should report termination message from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:247 - STEP: create the container 12/14/22 15:59:31.396 - STEP: wait for the container to reach Succeeded 12/14/22 15:59:31.409 - STEP: get the container status 12/14/22 15:59:35.459 - STEP: the container should be terminated 12/14/22 15:59:35.466 - STEP: the termination message should be set 12/14/22 15:59:35.466 - Dec 14 15:59:35.466: INFO: Expected: &{OK} to match Container's Termination Message: OK -- - STEP: delete the container 12/14/22 15:59:35.466 - [AfterEach] [sig-node] Container Runtime + STEP: Creating a kubernetes client 02/25/23 10:01:28.235 + Feb 25 10:01:28.235: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts 02/25/23 10:01:28.238 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:28.269 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:28.274 + [It] should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/kubelet_etc_hosts.go:63 + STEP: Setting up the test 02/25/23 10:01:28.279 + STEP: Creating hostNetwork=false pod 02/25/23 10:01:28.279 + Feb 25 10:01:28.295: INFO: Waiting up to 5m0s for pod "test-pod" in namespace "e2e-kubelet-etc-hosts-7469" to be "running and ready" + Feb 25 10:01:28.303: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 7.653424ms + Feb 25 10:01:28.303: INFO: The phase of Pod test-pod is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:01:30.310: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015230542s + Feb 25 10:01:30.311: INFO: The phase of Pod test-pod is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:01:32.309: INFO: Pod "test-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.014098589s + Feb 25 10:01:32.309: INFO: The phase of Pod test-pod is Running (Ready = true) + Feb 25 10:01:32.310: INFO: Pod "test-pod" satisfied condition "running and ready" + STEP: Creating hostNetwork=true pod 02/25/23 10:01:32.314 + Feb 25 10:01:32.323: INFO: Waiting up to 5m0s for pod "test-host-network-pod" in namespace "e2e-kubelet-etc-hosts-7469" to be "running and ready" + Feb 25 10:01:32.347: INFO: Pod "test-host-network-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 23.46451ms + Feb 25 10:01:32.347: INFO: The phase of Pod test-host-network-pod is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:01:34.354: INFO: Pod "test-host-network-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.031042507s + Feb 25 10:01:34.355: INFO: The phase of Pod test-host-network-pod is Running (Ready = true) + Feb 25 10:01:34.355: INFO: Pod "test-host-network-pod" satisfied condition "running and ready" + STEP: Running the test 02/25/23 10:01:34.362 + STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false 02/25/23 10:01:34.362 + Feb 25 10:01:34.363: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:01:34.363: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:01:34.367: INFO: ExecWithOptions: Clientset creation + Feb 25 10:01:34.368: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-1&container=busybox-1&stderr=true&stdout=true) + Feb 25 10:01:34.494: INFO: Exec stderr: "" + Feb 25 10:01:34.495: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:01:34.495: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:01:34.498: INFO: ExecWithOptions: Clientset creation + Feb 25 10:01:34.498: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-1&container=busybox-1&stderr=true&stdout=true) + Feb 25 10:01:34.580: INFO: Exec stderr: "" + Feb 25 10:01:34.580: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:01:34.581: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:01:34.583: INFO: ExecWithOptions: Clientset creation + Feb 25 10:01:34.583: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-2&container=busybox-2&stderr=true&stdout=true) + Feb 25 10:01:34.687: INFO: Exec stderr: "" + Feb 25 10:01:34.687: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:01:34.687: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:01:34.690: INFO: ExecWithOptions: Clientset creation + Feb 25 10:01:34.690: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-2&container=busybox-2&stderr=true&stdout=true) + Feb 25 10:01:34.770: INFO: Exec stderr: "" + STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount 02/25/23 10:01:34.77 + Feb 25 10:01:34.771: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:01:34.771: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:01:34.776: INFO: ExecWithOptions: Clientset creation + Feb 25 10:01:34.776: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-3&container=busybox-3&stderr=true&stdout=true) + Feb 25 10:01:34.917: INFO: Exec stderr: "" + Feb 25 10:01:34.917: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:01:34.917: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:01:34.920: INFO: ExecWithOptions: Clientset creation + Feb 25 10:01:34.920: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-3&container=busybox-3&stderr=true&stdout=true) + Feb 25 10:01:35.008: INFO: Exec stderr: "" + STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true 02/25/23 10:01:35.008 + Feb 25 10:01:35.008: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:01:35.008: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:01:35.010: INFO: ExecWithOptions: Clientset creation + Feb 25 10:01:35.010: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-1&container=busybox-1&stderr=true&stdout=true) + Feb 25 10:01:35.145: INFO: Exec stderr: "" + Feb 25 10:01:35.145: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:01:35.145: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:01:35.146: INFO: ExecWithOptions: Clientset creation + Feb 25 10:01:35.146: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-1&container=busybox-1&stderr=true&stdout=true) + Feb 25 10:01:35.284: INFO: Exec stderr: "" + Feb 25 10:01:35.284: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:01:35.285: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:01:35.287: INFO: ExecWithOptions: Clientset creation + Feb 25 10:01:35.287: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-2&container=busybox-2&stderr=true&stdout=true) + Feb 25 10:01:35.385: INFO: Exec stderr: "" + Feb 25 10:01:35.386: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-7469 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:01:35.386: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:01:35.388: INFO: ExecWithOptions: Clientset creation + Feb 25 10:01:35.388: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-7469/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-2&container=busybox-2&stderr=true&stdout=true) + Feb 25 10:01:35.477: INFO: Exec stderr: "" + [AfterEach] [sig-node] KubeletManagedEtcHosts test/e2e/framework/framework.go:187 - Dec 14 15:59:35.490: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-runtime-4074" for this suite. 12/14/22 15:59:35.495 + Feb 25 10:01:35.477: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "e2e-kubelet-etc-hosts-7469" for this suite. 02/25/23 10:01:35.486 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Pods Extended Pods Set QOS Class - should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] - test/e2e/node/pods.go:161 -[BeforeEach] [sig-node] Pods Extended +[sig-network] DNS + should support configurable pod DNS nameservers [Conformance] + test/e2e/network/dns.go:411 +[BeforeEach] [sig-network] DNS test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:59:35.502 -Dec 14 15:59:35.503: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename pods 12/14/22 15:59:35.504 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:35.527 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:35.531 -[BeforeEach] Pods Set QOS Class - test/e2e/node/pods.go:152 -[It] should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] - test/e2e/node/pods.go:161 -STEP: creating the pod 12/14/22 15:59:35.534 -STEP: submitting the pod to kubernetes 12/14/22 15:59:35.535 -STEP: verifying QOS class is set on the pod 12/14/22 15:59:35.548 -[AfterEach] [sig-node] Pods Extended +STEP: Creating a kubernetes client 02/25/23 10:01:35.522 +Feb 25 10:01:35.523: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename dns 02/25/23 10:01:35.525 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:35.552 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:35.557 +[It] should support configurable pod DNS nameservers [Conformance] + test/e2e/network/dns.go:411 +STEP: Creating a pod with dnsPolicy=None and customized dnsConfig... 02/25/23 10:01:35.568 +Feb 25 10:01:35.586: INFO: Created pod &Pod{ObjectMeta:{test-dns-nameservers dns-987 2aef0997-c0e2-4cb9-a90d-5255813a610e 13408 0 2023-02-25 10:01:35 +0000 UTC map[] map[] [] [] [{e2e.test Update v1 2023-02-25 10:01:35 +0000 UTC FieldsV1 {"f:spec":{"f:containers":{"k:{\"name\":\"agnhost-container\"}":{".":{},"f:args":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsConfig":{".":{},"f:nameservers":{},"f:searches":{}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-pz5xs,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost-container,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,Command:[],Args:[pause],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pz5xs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:None,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:&PodDNSConfig{Nameservers:[1.1.1.1],Searches:[resolv.conf.local],Options:[]PodDNSConfigOption{},},ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 10:01:35.587: INFO: Waiting up to 5m0s for pod "test-dns-nameservers" in namespace "dns-987" to be "running and ready" +Feb 25 10:01:35.595: INFO: Pod "test-dns-nameservers": Phase="Pending", Reason="", readiness=false. Elapsed: 7.275795ms +Feb 25 10:01:35.595: INFO: The phase of Pod test-dns-nameservers is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:01:37.603: INFO: Pod "test-dns-nameservers": Phase="Running", Reason="", readiness=true. Elapsed: 2.016002182s +Feb 25 10:01:37.604: INFO: The phase of Pod test-dns-nameservers is Running (Ready = true) +Feb 25 10:01:37.604: INFO: Pod "test-dns-nameservers" satisfied condition "running and ready" +STEP: Verifying customized DNS suffix list is configured on pod... 02/25/23 10:01:37.604 +Feb 25 10:01:37.605: INFO: ExecWithOptions {Command:[/agnhost dns-suffix] Namespace:dns-987 PodName:test-dns-nameservers ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:01:37.605: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:01:37.607: INFO: ExecWithOptions: Clientset creation +Feb 25 10:01:37.607: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/dns-987/pods/test-dns-nameservers/exec?command=%2Fagnhost&command=dns-suffix&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) +STEP: Verifying customized DNS server is configured on pod... 02/25/23 10:01:37.811 +Feb 25 10:01:37.811: INFO: ExecWithOptions {Command:[/agnhost dns-server-list] Namespace:dns-987 PodName:test-dns-nameservers ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:01:37.812: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:01:37.813: INFO: ExecWithOptions: Clientset creation +Feb 25 10:01:37.813: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/dns-987/pods/test-dns-nameservers/exec?command=%2Fagnhost&command=dns-server-list&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) +Feb 25 10:01:37.949: INFO: Deleting pod test-dns-nameservers... +[AfterEach] [sig-network] DNS test/e2e/framework/framework.go:187 -Dec 14 15:59:35.552: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-7848" for this suite. 12/14/22 15:59:35.56 -{"msg":"PASSED [sig-node] Pods Extended Pods Set QOS Class should be set on Pods with matching resource requests and limits for memory and cpu [Conformance]","completed":62,"skipped":1126,"failed":0} +Feb 25 10:01:37.973: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "dns-987" for this suite. 02/25/23 10:01:37.982 +{"msg":"PASSED [sig-network] DNS should support configurable pod DNS nameservers [Conformance]","completed":62,"skipped":1266,"failed":0} ------------------------------ -• [0.075 seconds] -[sig-node] Pods Extended -test/e2e/node/framework.go:23 - Pods Set QOS Class - test/e2e/node/pods.go:150 - should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] - test/e2e/node/pods.go:161 +• [2.478 seconds] +[sig-network] DNS +test/e2e/network/common/framework.go:23 + should support configurable pod DNS nameservers [Conformance] + test/e2e/network/dns.go:411 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Pods Extended + [BeforeEach] [sig-network] DNS test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:59:35.502 - Dec 14 15:59:35.503: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename pods 12/14/22 15:59:35.504 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:35.527 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:35.531 - [BeforeEach] Pods Set QOS Class - test/e2e/node/pods.go:152 - [It] should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] - test/e2e/node/pods.go:161 - STEP: creating the pod 12/14/22 15:59:35.534 - STEP: submitting the pod to kubernetes 12/14/22 15:59:35.535 - STEP: verifying QOS class is set on the pod 12/14/22 15:59:35.548 - [AfterEach] [sig-node] Pods Extended + STEP: Creating a kubernetes client 02/25/23 10:01:35.522 + Feb 25 10:01:35.523: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename dns 02/25/23 10:01:35.525 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:35.552 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:35.557 + [It] should support configurable pod DNS nameservers [Conformance] + test/e2e/network/dns.go:411 + STEP: Creating a pod with dnsPolicy=None and customized dnsConfig... 02/25/23 10:01:35.568 + Feb 25 10:01:35.586: INFO: Created pod &Pod{ObjectMeta:{test-dns-nameservers dns-987 2aef0997-c0e2-4cb9-a90d-5255813a610e 13408 0 2023-02-25 10:01:35 +0000 UTC map[] map[] [] [] [{e2e.test Update v1 2023-02-25 10:01:35 +0000 UTC FieldsV1 {"f:spec":{"f:containers":{"k:{\"name\":\"agnhost-container\"}":{".":{},"f:args":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsConfig":{".":{},"f:nameservers":{},"f:searches":{}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-pz5xs,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost-container,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,Command:[],Args:[pause],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pz5xs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:None,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:&PodDNSConfig{Nameservers:[1.1.1.1],Searches:[resolv.conf.local],Options:[]PodDNSConfigOption{},},ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 10:01:35.587: INFO: Waiting up to 5m0s for pod "test-dns-nameservers" in namespace "dns-987" to be "running and ready" + Feb 25 10:01:35.595: INFO: Pod "test-dns-nameservers": Phase="Pending", Reason="", readiness=false. Elapsed: 7.275795ms + Feb 25 10:01:35.595: INFO: The phase of Pod test-dns-nameservers is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:01:37.603: INFO: Pod "test-dns-nameservers": Phase="Running", Reason="", readiness=true. Elapsed: 2.016002182s + Feb 25 10:01:37.604: INFO: The phase of Pod test-dns-nameservers is Running (Ready = true) + Feb 25 10:01:37.604: INFO: Pod "test-dns-nameservers" satisfied condition "running and ready" + STEP: Verifying customized DNS suffix list is configured on pod... 02/25/23 10:01:37.604 + Feb 25 10:01:37.605: INFO: ExecWithOptions {Command:[/agnhost dns-suffix] Namespace:dns-987 PodName:test-dns-nameservers ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:01:37.605: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:01:37.607: INFO: ExecWithOptions: Clientset creation + Feb 25 10:01:37.607: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/dns-987/pods/test-dns-nameservers/exec?command=%2Fagnhost&command=dns-suffix&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) + STEP: Verifying customized DNS server is configured on pod... 02/25/23 10:01:37.811 + Feb 25 10:01:37.811: INFO: ExecWithOptions {Command:[/agnhost dns-server-list] Namespace:dns-987 PodName:test-dns-nameservers ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:01:37.812: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:01:37.813: INFO: ExecWithOptions: Clientset creation + Feb 25 10:01:37.813: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/dns-987/pods/test-dns-nameservers/exec?command=%2Fagnhost&command=dns-server-list&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) + Feb 25 10:01:37.949: INFO: Deleting pod test-dns-nameservers... + [AfterEach] [sig-network] DNS test/e2e/framework/framework.go:187 - Dec 14 15:59:35.552: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "pods-7848" for this suite. 12/14/22 15:59:35.56 + Feb 25 10:01:37.973: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "dns-987" for this suite. 02/25/23 10:01:37.982 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] Watchers - should be able to start watching from a specific resource version [Conformance] - test/e2e/apimachinery/watch.go:142 -[BeforeEach] [sig-api-machinery] Watchers +[sig-apps] ReplicaSet + should serve a basic image on each replica with a public image [Conformance] + test/e2e/apps/replica_set.go:111 +[BeforeEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:59:35.583 -Dec 14 15:59:35.583: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename watch 12/14/22 15:59:35.586 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:35.611 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:35.615 -[It] should be able to start watching from a specific resource version [Conformance] - test/e2e/apimachinery/watch.go:142 -STEP: creating a new configmap 12/14/22 15:59:35.618 -STEP: modifying the configmap once 12/14/22 15:59:35.627 -STEP: modifying the configmap a second time 12/14/22 15:59:35.638 -STEP: deleting the configmap 12/14/22 15:59:35.649 -STEP: creating a watch on configmaps from the resource version returned by the first update 12/14/22 15:59:35.658 -STEP: Expecting to observe notifications for all changes to the configmap after the first update 12/14/22 15:59:35.659 -Dec 14 15:59:35.660: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-8032 50fca8cd-073e-4b88-b1df-8941ac572566 9891 0 2022-12-14 15:59:35 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] [{e2e.test Update v1 2022-12-14 15:59:35 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} -Dec 14 15:59:35.661: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-8032 50fca8cd-073e-4b88-b1df-8941ac572566 9892 0 2022-12-14 15:59:35 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] [{e2e.test Update v1 2022-12-14 15:59:35 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} -[AfterEach] [sig-api-machinery] Watchers +STEP: Creating a kubernetes client 02/25/23 10:01:38.003 +Feb 25 10:01:38.003: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename replicaset 02/25/23 10:01:38.005 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:38.042 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:38.045 +[It] should serve a basic image on each replica with a public image [Conformance] + test/e2e/apps/replica_set.go:111 +Feb 25 10:01:38.050: INFO: Creating ReplicaSet my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d +Feb 25 10:01:38.067: INFO: Pod name my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d: Found 0 pods out of 1 +Feb 25 10:01:43.078: INFO: Pod name my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d: Found 1 pods out of 1 +Feb 25 10:01:43.078: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d" is running +Feb 25 10:01:43.078: INFO: Waiting up to 5m0s for pod "my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d-xrwr4" in namespace "replicaset-8226" to be "running" +Feb 25 10:01:43.086: INFO: Pod "my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d-xrwr4": Phase="Running", Reason="", readiness=true. Elapsed: 8.011747ms +Feb 25 10:01:43.087: INFO: Pod "my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d-xrwr4" satisfied condition "running" +Feb 25 10:01:43.087: INFO: Pod "my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d-xrwr4" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:01:38 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:01:39 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:01:39 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:01:38 +0000 UTC Reason: Message:}]) +Feb 25 10:01:43.087: INFO: Trying to dial the pod +Feb 25 10:01:48.114: INFO: Controller my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d: Got expected result from replica 1 [my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d-xrwr4]: "my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d-xrwr4", 1 of 1 required successes so far +[AfterEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:187 -Dec 14 15:59:35.661: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "watch-8032" for this suite. 12/14/22 15:59:35.666 -{"msg":"PASSED [sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance]","completed":63,"skipped":1137,"failed":0} +Feb 25 10:01:48.114: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "replicaset-8226" for this suite. 02/25/23 10:01:48.124 +{"msg":"PASSED [sig-apps] ReplicaSet should serve a basic image on each replica with a public image [Conformance]","completed":63,"skipped":1300,"failed":0} ------------------------------ -• [0.092 seconds] -[sig-api-machinery] Watchers -test/e2e/apimachinery/framework.go:23 - should be able to start watching from a specific resource version [Conformance] - test/e2e/apimachinery/watch.go:142 +• [SLOW TEST] [10.132 seconds] +[sig-apps] ReplicaSet +test/e2e/apps/framework.go:23 + should serve a basic image on each replica with a public image [Conformance] + test/e2e/apps/replica_set.go:111 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Watchers + [BeforeEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:59:35.583 - Dec 14 15:59:35.583: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename watch 12/14/22 15:59:35.586 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:35.611 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:35.615 - [It] should be able to start watching from a specific resource version [Conformance] - test/e2e/apimachinery/watch.go:142 - STEP: creating a new configmap 12/14/22 15:59:35.618 - STEP: modifying the configmap once 12/14/22 15:59:35.627 - STEP: modifying the configmap a second time 12/14/22 15:59:35.638 - STEP: deleting the configmap 12/14/22 15:59:35.649 - STEP: creating a watch on configmaps from the resource version returned by the first update 12/14/22 15:59:35.658 - STEP: Expecting to observe notifications for all changes to the configmap after the first update 12/14/22 15:59:35.659 - Dec 14 15:59:35.660: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-8032 50fca8cd-073e-4b88-b1df-8941ac572566 9891 0 2022-12-14 15:59:35 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] [{e2e.test Update v1 2022-12-14 15:59:35 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} - Dec 14 15:59:35.661: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-8032 50fca8cd-073e-4b88-b1df-8941ac572566 9892 0 2022-12-14 15:59:35 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] [{e2e.test Update v1 2022-12-14 15:59:35 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} - [AfterEach] [sig-api-machinery] Watchers + STEP: Creating a kubernetes client 02/25/23 10:01:38.003 + Feb 25 10:01:38.003: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename replicaset 02/25/23 10:01:38.005 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:38.042 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:38.045 + [It] should serve a basic image on each replica with a public image [Conformance] + test/e2e/apps/replica_set.go:111 + Feb 25 10:01:38.050: INFO: Creating ReplicaSet my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d + Feb 25 10:01:38.067: INFO: Pod name my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d: Found 0 pods out of 1 + Feb 25 10:01:43.078: INFO: Pod name my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d: Found 1 pods out of 1 + Feb 25 10:01:43.078: INFO: Ensuring a pod for ReplicaSet "my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d" is running + Feb 25 10:01:43.078: INFO: Waiting up to 5m0s for pod "my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d-xrwr4" in namespace "replicaset-8226" to be "running" + Feb 25 10:01:43.086: INFO: Pod "my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d-xrwr4": Phase="Running", Reason="", readiness=true. Elapsed: 8.011747ms + Feb 25 10:01:43.087: INFO: Pod "my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d-xrwr4" satisfied condition "running" + Feb 25 10:01:43.087: INFO: Pod "my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d-xrwr4" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:01:38 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:01:39 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:01:39 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:01:38 +0000 UTC Reason: Message:}]) + Feb 25 10:01:43.087: INFO: Trying to dial the pod + Feb 25 10:01:48.114: INFO: Controller my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d: Got expected result from replica 1 [my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d-xrwr4]: "my-hostname-basic-92c519f1-5b80-4246-836f-2b21519f5f2d-xrwr4", 1 of 1 required successes so far + [AfterEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:187 - Dec 14 15:59:35.661: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "watch-8032" for this suite. 12/14/22 15:59:35.666 + Feb 25 10:01:48.114: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "replicaset-8226" for this suite. 02/25/23 10:01:48.124 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSS +SSSSSSS ------------------------------ [sig-storage] Projected downwardAPI - should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:83 + should provide container's cpu request [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:220 [BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:59:35.68 -Dec 14 15:59:35.680: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 15:59:35.682 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:35.705 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:35.709 +STEP: Creating a kubernetes client 02/25/23 10:01:48.137 +Feb 25 10:01:48.137: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 10:01:48.138 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:48.169 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:48.174 [BeforeEach] [sig-storage] Projected downwardAPI test/e2e/common/storage/projected_downwardapi.go:43 -[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:83 -STEP: Creating a pod to test downward API volume plugin 12/14/22 15:59:35.713 -Dec 14 15:59:35.728: INFO: Waiting up to 5m0s for pod "downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac" in namespace "projected-5222" to be "Succeeded or Failed" -Dec 14 15:59:35.732: INFO: Pod "downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac": Phase="Pending", Reason="", readiness=false. Elapsed: 3.622573ms -Dec 14 15:59:37.739: INFO: Pod "downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011352348s -Dec 14 15:59:39.743: INFO: Pod "downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015028485s -STEP: Saw pod success 12/14/22 15:59:39.743 -Dec 14 15:59:39.745: INFO: Pod "downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac" satisfied condition "Succeeded or Failed" -Dec 14 15:59:39.751: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac container client-container: -STEP: delete the pod 12/14/22 15:59:39.861 -Dec 14 15:59:39.877: INFO: Waiting for pod downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac to disappear -Dec 14 15:59:39.881: INFO: Pod downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac no longer exists +[It] should provide container's cpu request [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:220 +STEP: Creating a pod to test downward API volume plugin 02/25/23 10:01:48.179 +Feb 25 10:01:48.200: INFO: Waiting up to 5m0s for pod "downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f" in namespace "projected-6216" to be "Succeeded or Failed" +Feb 25 10:01:48.207: INFO: Pod "downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f": Phase="Pending", Reason="", readiness=false. Elapsed: 7.678956ms +Feb 25 10:01:50.214: INFO: Pod "downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014441121s +Feb 25 10:01:52.215: INFO: Pod "downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015267236s +STEP: Saw pod success 02/25/23 10:01:52.215 +Feb 25 10:01:52.215: INFO: Pod "downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f" satisfied condition "Succeeded or Failed" +Feb 25 10:01:52.223: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f container client-container: +STEP: delete the pod 02/25/23 10:01:52.235 +Feb 25 10:01:52.257: INFO: Waiting for pod downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f to disappear +Feb 25 10:01:52.262: INFO: Pod downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f no longer exists [AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 -Dec 14 15:59:39.881: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-5222" for this suite. 12/14/22 15:59:39.887 -{"msg":"PASSED [sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","completed":64,"skipped":1156,"failed":0} +Feb 25 10:01:52.263: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-6216" for this suite. 02/25/23 10:01:52.272 +{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu request [NodeConformance] [Conformance]","completed":64,"skipped":1307,"failed":0} ------------------------------ -• [4.214 seconds] +• [4.148 seconds] [sig-storage] Projected downwardAPI test/e2e/common/storage/framework.go:23 - should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:83 + should provide container's cpu request [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:220 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:59:35.68 - Dec 14 15:59:35.680: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 15:59:35.682 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:35.705 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:35.709 + STEP: Creating a kubernetes client 02/25/23 10:01:48.137 + Feb 25 10:01:48.137: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 10:01:48.138 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:48.169 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:48.174 [BeforeEach] [sig-storage] Projected downwardAPI test/e2e/common/storage/projected_downwardapi.go:43 - [It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:83 - STEP: Creating a pod to test downward API volume plugin 12/14/22 15:59:35.713 - Dec 14 15:59:35.728: INFO: Waiting up to 5m0s for pod "downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac" in namespace "projected-5222" to be "Succeeded or Failed" - Dec 14 15:59:35.732: INFO: Pod "downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac": Phase="Pending", Reason="", readiness=false. Elapsed: 3.622573ms - Dec 14 15:59:37.739: INFO: Pod "downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011352348s - Dec 14 15:59:39.743: INFO: Pod "downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015028485s - STEP: Saw pod success 12/14/22 15:59:39.743 - Dec 14 15:59:39.745: INFO: Pod "downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac" satisfied condition "Succeeded or Failed" - Dec 14 15:59:39.751: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac container client-container: - STEP: delete the pod 12/14/22 15:59:39.861 - Dec 14 15:59:39.877: INFO: Waiting for pod downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac to disappear - Dec 14 15:59:39.881: INFO: Pod downwardapi-volume-a7a1a8e9-1a2d-43e1-aaa6-88a33eb861ac no longer exists + [It] should provide container's cpu request [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:220 + STEP: Creating a pod to test downward API volume plugin 02/25/23 10:01:48.179 + Feb 25 10:01:48.200: INFO: Waiting up to 5m0s for pod "downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f" in namespace "projected-6216" to be "Succeeded or Failed" + Feb 25 10:01:48.207: INFO: Pod "downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f": Phase="Pending", Reason="", readiness=false. Elapsed: 7.678956ms + Feb 25 10:01:50.214: INFO: Pod "downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014441121s + Feb 25 10:01:52.215: INFO: Pod "downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015267236s + STEP: Saw pod success 02/25/23 10:01:52.215 + Feb 25 10:01:52.215: INFO: Pod "downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f" satisfied condition "Succeeded or Failed" + Feb 25 10:01:52.223: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f container client-container: + STEP: delete the pod 02/25/23 10:01:52.235 + Feb 25 10:01:52.257: INFO: Waiting for pod downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f to disappear + Feb 25 10:01:52.262: INFO: Pod downwardapi-volume-fbd404e0-e190-4054-bc3a-ec301e69c49f no longer exists [AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 - Dec 14 15:59:39.881: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-5222" for this suite. 12/14/22 15:59:39.887 + Feb 25 10:01:52.263: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-6216" for this suite. 02/25/23 10:01:52.272 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Downward API - should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:43 -[BeforeEach] [sig-node] Downward API +[sig-storage] Downward API volume + should update labels on modification [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:129 +[BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:59:39.9 -Dec 14 15:59:39.900: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 15:59:39.902 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:39.933 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:39.937 -[It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:43 -STEP: Creating a pod to test downward api env vars 12/14/22 15:59:39.941 -Dec 14 15:59:39.957: INFO: Waiting up to 5m0s for pod "downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94" in namespace "downward-api-1833" to be "Succeeded or Failed" -Dec 14 15:59:39.963: INFO: Pod "downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94": Phase="Pending", Reason="", readiness=false. Elapsed: 5.941291ms -Dec 14 15:59:41.971: INFO: Pod "downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01397623s -Dec 14 15:59:43.970: INFO: Pod "downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94": Phase="Pending", Reason="", readiness=false. Elapsed: 4.013050832s -Dec 14 15:59:45.970: INFO: Pod "downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.012704126s -STEP: Saw pod success 12/14/22 15:59:45.97 -Dec 14 15:59:45.970: INFO: Pod "downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94" satisfied condition "Succeeded or Failed" -Dec 14 15:59:45.975: INFO: Trying to get logs from node kiedai4iakae-3 pod downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94 container dapi-container: -STEP: delete the pod 12/14/22 15:59:45.987 -Dec 14 15:59:46.023: INFO: Waiting for pod downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94 to disappear -Dec 14 15:59:46.032: INFO: Pod downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94 no longer exists -[AfterEach] [sig-node] Downward API +STEP: Creating a kubernetes client 02/25/23 10:01:52.295 +Feb 25 10:01:52.296: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 10:01:52.297 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:52.325 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:52.332 +[BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 +[It] should update labels on modification [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:129 +STEP: Creating the pod 02/25/23 10:01:52.336 +Feb 25 10:01:52.350: INFO: Waiting up to 5m0s for pod "labelsupdatee648334d-ac61-42a6-bcac-d7f5dae12547" in namespace "downward-api-3355" to be "running and ready" +Feb 25 10:01:52.360: INFO: Pod "labelsupdatee648334d-ac61-42a6-bcac-d7f5dae12547": Phase="Pending", Reason="", readiness=false. Elapsed: 9.284568ms +Feb 25 10:01:52.360: INFO: The phase of Pod labelsupdatee648334d-ac61-42a6-bcac-d7f5dae12547 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:01:54.370: INFO: Pod "labelsupdatee648334d-ac61-42a6-bcac-d7f5dae12547": Phase="Running", Reason="", readiness=true. Elapsed: 2.020022657s +Feb 25 10:01:54.371: INFO: The phase of Pod labelsupdatee648334d-ac61-42a6-bcac-d7f5dae12547 is Running (Ready = true) +Feb 25 10:01:54.371: INFO: Pod "labelsupdatee648334d-ac61-42a6-bcac-d7f5dae12547" satisfied condition "running and ready" +Feb 25 10:01:54.921: INFO: Successfully updated pod "labelsupdatee648334d-ac61-42a6-bcac-d7f5dae12547" +[AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 -Dec 14 15:59:46.032: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-1833" for this suite. 12/14/22 15:59:46.057 -{"msg":"PASSED [sig-node] Downward API should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance]","completed":65,"skipped":1164,"failed":0} +Feb 25 10:01:59.123: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-3355" for this suite. 02/25/23 10:01:59.132 +{"msg":"PASSED [sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance]","completed":65,"skipped":1349,"failed":0} ------------------------------ -• [SLOW TEST] [6.174 seconds] -[sig-node] Downward API -test/e2e/common/node/framework.go:23 - should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:43 +• [SLOW TEST] [6.849 seconds] +[sig-storage] Downward API volume +test/e2e/common/storage/framework.go:23 + should update labels on modification [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:129 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Downward API + [BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:59:39.9 - Dec 14 15:59:39.900: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 15:59:39.902 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:39.933 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:39.937 - [It] should provide pod name, namespace and IP address as env vars [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:43 - STEP: Creating a pod to test downward api env vars 12/14/22 15:59:39.941 - Dec 14 15:59:39.957: INFO: Waiting up to 5m0s for pod "downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94" in namespace "downward-api-1833" to be "Succeeded or Failed" - Dec 14 15:59:39.963: INFO: Pod "downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94": Phase="Pending", Reason="", readiness=false. Elapsed: 5.941291ms - Dec 14 15:59:41.971: INFO: Pod "downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01397623s - Dec 14 15:59:43.970: INFO: Pod "downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94": Phase="Pending", Reason="", readiness=false. Elapsed: 4.013050832s - Dec 14 15:59:45.970: INFO: Pod "downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.012704126s - STEP: Saw pod success 12/14/22 15:59:45.97 - Dec 14 15:59:45.970: INFO: Pod "downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94" satisfied condition "Succeeded or Failed" - Dec 14 15:59:45.975: INFO: Trying to get logs from node kiedai4iakae-3 pod downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94 container dapi-container: - STEP: delete the pod 12/14/22 15:59:45.987 - Dec 14 15:59:46.023: INFO: Waiting for pod downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94 to disappear - Dec 14 15:59:46.032: INFO: Pod downward-api-f37b70c4-d72e-49b8-b5e0-737473a07d94 no longer exists - [AfterEach] [sig-node] Downward API + STEP: Creating a kubernetes client 02/25/23 10:01:52.295 + Feb 25 10:01:52.296: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 10:01:52.297 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:52.325 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:52.332 + [BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 + [It] should update labels on modification [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:129 + STEP: Creating the pod 02/25/23 10:01:52.336 + Feb 25 10:01:52.350: INFO: Waiting up to 5m0s for pod "labelsupdatee648334d-ac61-42a6-bcac-d7f5dae12547" in namespace "downward-api-3355" to be "running and ready" + Feb 25 10:01:52.360: INFO: Pod "labelsupdatee648334d-ac61-42a6-bcac-d7f5dae12547": Phase="Pending", Reason="", readiness=false. Elapsed: 9.284568ms + Feb 25 10:01:52.360: INFO: The phase of Pod labelsupdatee648334d-ac61-42a6-bcac-d7f5dae12547 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:01:54.370: INFO: Pod "labelsupdatee648334d-ac61-42a6-bcac-d7f5dae12547": Phase="Running", Reason="", readiness=true. Elapsed: 2.020022657s + Feb 25 10:01:54.371: INFO: The phase of Pod labelsupdatee648334d-ac61-42a6-bcac-d7f5dae12547 is Running (Ready = true) + Feb 25 10:01:54.371: INFO: Pod "labelsupdatee648334d-ac61-42a6-bcac-d7f5dae12547" satisfied condition "running and ready" + Feb 25 10:01:54.921: INFO: Successfully updated pod "labelsupdatee648334d-ac61-42a6-bcac-d7f5dae12547" + [AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 - Dec 14 15:59:46.032: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-1833" for this suite. 12/14/22 15:59:46.057 + Feb 25 10:01:59.123: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-3355" for this suite. 02/25/23 10:01:59.132 << End Captured GinkgoWriter Output ------------------------------ -SS ------------------------------- -[sig-node] Pods - should support remote command execution over websockets [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:535 -[BeforeEach] [sig-node] Pods +[sig-apps] Daemon set [Serial] + should verify changes to a daemon set status [Conformance] + test/e2e/apps/daemon_set.go:861 +[BeforeEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:59:46.075 -Dec 14 15:59:46.075: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename pods 12/14/22 15:59:46.077 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:46.103 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:46.108 -[BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 -[It] should support remote command execution over websockets [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:535 -Dec 14 15:59:46.112: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: creating the pod 12/14/22 15:59:46.113 -STEP: submitting the pod to kubernetes 12/14/22 15:59:46.114 -Dec 14 15:59:46.138: INFO: Waiting up to 5m0s for pod "pod-exec-websocket-46433553-04d6-4355-845b-411c4cbcda17" in namespace "pods-5133" to be "running and ready" -Dec 14 15:59:46.143: INFO: Pod "pod-exec-websocket-46433553-04d6-4355-845b-411c4cbcda17": Phase="Pending", Reason="", readiness=false. Elapsed: 4.700182ms -Dec 14 15:59:46.143: INFO: The phase of Pod pod-exec-websocket-46433553-04d6-4355-845b-411c4cbcda17 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 15:59:48.149: INFO: Pod "pod-exec-websocket-46433553-04d6-4355-845b-411c4cbcda17": Phase="Running", Reason="", readiness=true. Elapsed: 2.011032601s -Dec 14 15:59:48.149: INFO: The phase of Pod pod-exec-websocket-46433553-04d6-4355-845b-411c4cbcda17 is Running (Ready = true) -Dec 14 15:59:48.149: INFO: Pod "pod-exec-websocket-46433553-04d6-4355-845b-411c4cbcda17" satisfied condition "running and ready" -[AfterEach] [sig-node] Pods +STEP: Creating a kubernetes client 02/25/23 10:01:59.147 +Feb 25 10:01:59.147: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename daemonsets 02/25/23 10:01:59.15 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:59.182 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:59.19 +[BeforeEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:145 +[It] should verify changes to a daemon set status [Conformance] + test/e2e/apps/daemon_set.go:861 +STEP: Creating simple DaemonSet "daemon-set" 02/25/23 10:01:59.249 +STEP: Check that daemon pods launch on every node of the cluster. 02/25/23 10:01:59.264 +Feb 25 10:01:59.283: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:01:59.283: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 10:02:00.321: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:02:00.321: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 10:02:01.302: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 +Feb 25 10:02:01.302: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set +STEP: Getting /status 02/25/23 10:02:01.308 +Feb 25 10:02:01.316: INFO: Daemon Set daemon-set has Conditions: [] +STEP: updating the DaemonSet Status 02/25/23 10:02:01.316 +Feb 25 10:02:01.333: INFO: updatedStatus.Conditions: []v1.DaemonSetCondition{v1.DaemonSetCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} +STEP: watching for the daemon set status to be updated 02/25/23 10:02:01.334 +Feb 25 10:02:01.339: INFO: Observed &DaemonSet event: ADDED +Feb 25 10:02:01.339: INFO: Observed &DaemonSet event: MODIFIED +Feb 25 10:02:01.340: INFO: Observed &DaemonSet event: MODIFIED +Feb 25 10:02:01.340: INFO: Observed &DaemonSet event: MODIFIED +Feb 25 10:02:01.341: INFO: Observed &DaemonSet event: MODIFIED +Feb 25 10:02:01.341: INFO: Observed &DaemonSet event: MODIFIED +Feb 25 10:02:01.341: INFO: Found daemon set daemon-set in namespace daemonsets-118 with labels: map[daemonset-name:daemon-set] annotations: map[deprecated.daemonset.template.generation:1] & Conditions: [{StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] +Feb 25 10:02:01.342: INFO: Daemon set daemon-set has an updated status +STEP: patching the DaemonSet Status 02/25/23 10:02:01.342 +STEP: watching for the daemon set status to be patched 02/25/23 10:02:01.359 +Feb 25 10:02:01.365: INFO: Observed &DaemonSet event: ADDED +Feb 25 10:02:01.366: INFO: Observed &DaemonSet event: MODIFIED +Feb 25 10:02:01.367: INFO: Observed &DaemonSet event: MODIFIED +Feb 25 10:02:01.367: INFO: Observed &DaemonSet event: MODIFIED +Feb 25 10:02:01.368: INFO: Observed &DaemonSet event: MODIFIED +Feb 25 10:02:01.368: INFO: Observed &DaemonSet event: MODIFIED +Feb 25 10:02:01.369: INFO: Observed daemon set daemon-set in namespace daemonsets-118 with annotations: map[deprecated.daemonset.template.generation:1] & Conditions: [{StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] +Feb 25 10:02:01.369: INFO: Observed &DaemonSet event: MODIFIED +Feb 25 10:02:01.369: INFO: Found daemon set daemon-set in namespace daemonsets-118 with labels: map[daemonset-name:daemon-set] annotations: map[deprecated.daemonset.template.generation:1] & Conditions: [{StatusPatched True 0001-01-01 00:00:00 +0000 UTC }] +Feb 25 10:02:01.369: INFO: Daemon set daemon-set has a patched status +[AfterEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:110 +STEP: Deleting DaemonSet "daemon-set" 02/25/23 10:02:01.377 +STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-118, will wait for the garbage collector to delete the pods 02/25/23 10:02:01.377 +Feb 25 10:02:01.457: INFO: Deleting DaemonSet.extensions daemon-set took: 18.123516ms +Feb 25 10:02:01.558: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.948591ms +Feb 25 10:02:04.469: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:02:04.469: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set +Feb 25 10:02:04.475: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"13669"},"items":null} + +Feb 25 10:02:04.480: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"13669"},"items":null} + +[AfterEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:187 -Dec 14 15:59:48.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-5133" for this suite. 12/14/22 15:59:48.24 -{"msg":"PASSED [sig-node] Pods should support remote command execution over websockets [NodeConformance] [Conformance]","completed":66,"skipped":1166,"failed":0} +Feb 25 10:02:04.511: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "daemonsets-118" for this suite. 02/25/23 10:02:04.518 +{"msg":"PASSED [sig-apps] Daemon set [Serial] should verify changes to a daemon set status [Conformance]","completed":66,"skipped":1349,"failed":0} ------------------------------ -• [2.177 seconds] -[sig-node] Pods -test/e2e/common/node/framework.go:23 - should support remote command execution over websockets [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:535 +• [SLOW TEST] [5.382 seconds] +[sig-apps] Daemon set [Serial] +test/e2e/apps/framework.go:23 + should verify changes to a daemon set status [Conformance] + test/e2e/apps/daemon_set.go:861 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Pods + [BeforeEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:59:46.075 - Dec 14 15:59:46.075: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename pods 12/14/22 15:59:46.077 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:46.103 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:46.108 - [BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 - [It] should support remote command execution over websockets [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:535 - Dec 14 15:59:46.112: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: creating the pod 12/14/22 15:59:46.113 - STEP: submitting the pod to kubernetes 12/14/22 15:59:46.114 - Dec 14 15:59:46.138: INFO: Waiting up to 5m0s for pod "pod-exec-websocket-46433553-04d6-4355-845b-411c4cbcda17" in namespace "pods-5133" to be "running and ready" - Dec 14 15:59:46.143: INFO: Pod "pod-exec-websocket-46433553-04d6-4355-845b-411c4cbcda17": Phase="Pending", Reason="", readiness=false. Elapsed: 4.700182ms - Dec 14 15:59:46.143: INFO: The phase of Pod pod-exec-websocket-46433553-04d6-4355-845b-411c4cbcda17 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 15:59:48.149: INFO: Pod "pod-exec-websocket-46433553-04d6-4355-845b-411c4cbcda17": Phase="Running", Reason="", readiness=true. Elapsed: 2.011032601s - Dec 14 15:59:48.149: INFO: The phase of Pod pod-exec-websocket-46433553-04d6-4355-845b-411c4cbcda17 is Running (Ready = true) - Dec 14 15:59:48.149: INFO: Pod "pod-exec-websocket-46433553-04d6-4355-845b-411c4cbcda17" satisfied condition "running and ready" - [AfterEach] [sig-node] Pods + STEP: Creating a kubernetes client 02/25/23 10:01:59.147 + Feb 25 10:01:59.147: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename daemonsets 02/25/23 10:01:59.15 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:01:59.182 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:01:59.19 + [BeforeEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:145 + [It] should verify changes to a daemon set status [Conformance] + test/e2e/apps/daemon_set.go:861 + STEP: Creating simple DaemonSet "daemon-set" 02/25/23 10:01:59.249 + STEP: Check that daemon pods launch on every node of the cluster. 02/25/23 10:01:59.264 + Feb 25 10:01:59.283: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:01:59.283: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 10:02:00.321: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:02:00.321: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 10:02:01.302: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 + Feb 25 10:02:01.302: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set + STEP: Getting /status 02/25/23 10:02:01.308 + Feb 25 10:02:01.316: INFO: Daemon Set daemon-set has Conditions: [] + STEP: updating the DaemonSet Status 02/25/23 10:02:01.316 + Feb 25 10:02:01.333: INFO: updatedStatus.Conditions: []v1.DaemonSetCondition{v1.DaemonSetCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} + STEP: watching for the daemon set status to be updated 02/25/23 10:02:01.334 + Feb 25 10:02:01.339: INFO: Observed &DaemonSet event: ADDED + Feb 25 10:02:01.339: INFO: Observed &DaemonSet event: MODIFIED + Feb 25 10:02:01.340: INFO: Observed &DaemonSet event: MODIFIED + Feb 25 10:02:01.340: INFO: Observed &DaemonSet event: MODIFIED + Feb 25 10:02:01.341: INFO: Observed &DaemonSet event: MODIFIED + Feb 25 10:02:01.341: INFO: Observed &DaemonSet event: MODIFIED + Feb 25 10:02:01.341: INFO: Found daemon set daemon-set in namespace daemonsets-118 with labels: map[daemonset-name:daemon-set] annotations: map[deprecated.daemonset.template.generation:1] & Conditions: [{StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] + Feb 25 10:02:01.342: INFO: Daemon set daemon-set has an updated status + STEP: patching the DaemonSet Status 02/25/23 10:02:01.342 + STEP: watching for the daemon set status to be patched 02/25/23 10:02:01.359 + Feb 25 10:02:01.365: INFO: Observed &DaemonSet event: ADDED + Feb 25 10:02:01.366: INFO: Observed &DaemonSet event: MODIFIED + Feb 25 10:02:01.367: INFO: Observed &DaemonSet event: MODIFIED + Feb 25 10:02:01.367: INFO: Observed &DaemonSet event: MODIFIED + Feb 25 10:02:01.368: INFO: Observed &DaemonSet event: MODIFIED + Feb 25 10:02:01.368: INFO: Observed &DaemonSet event: MODIFIED + Feb 25 10:02:01.369: INFO: Observed daemon set daemon-set in namespace daemonsets-118 with annotations: map[deprecated.daemonset.template.generation:1] & Conditions: [{StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] + Feb 25 10:02:01.369: INFO: Observed &DaemonSet event: MODIFIED + Feb 25 10:02:01.369: INFO: Found daemon set daemon-set in namespace daemonsets-118 with labels: map[daemonset-name:daemon-set] annotations: map[deprecated.daemonset.template.generation:1] & Conditions: [{StatusPatched True 0001-01-01 00:00:00 +0000 UTC }] + Feb 25 10:02:01.369: INFO: Daemon set daemon-set has a patched status + [AfterEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:110 + STEP: Deleting DaemonSet "daemon-set" 02/25/23 10:02:01.377 + STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-118, will wait for the garbage collector to delete the pods 02/25/23 10:02:01.377 + Feb 25 10:02:01.457: INFO: Deleting DaemonSet.extensions daemon-set took: 18.123516ms + Feb 25 10:02:01.558: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.948591ms + Feb 25 10:02:04.469: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:02:04.469: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set + Feb 25 10:02:04.475: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"13669"},"items":null} + + Feb 25 10:02:04.480: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"13669"},"items":null} + + [AfterEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:187 - Dec 14 15:59:48.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "pods-5133" for this suite. 12/14/22 15:59:48.24 + Feb 25 10:02:04.511: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "daemonsets-118" for this suite. 02/25/23 10:02:04.518 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSS ------------------------------ -[sig-storage] Projected secret - should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:77 -[BeforeEach] [sig-storage] Projected secret +[sig-storage] EmptyDir volumes + should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:196 +[BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:59:48.258 -Dec 14 15:59:48.258: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 15:59:48.26 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:48.283 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:48.288 -[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:77 -STEP: Creating projection with secret that has name projected-secret-test-map-f811e5ad-cd33-45cb-83ef-f136f8cd2f11 12/14/22 15:59:48.292 -STEP: Creating a pod to test consume secrets 12/14/22 15:59:48.298 -Dec 14 15:59:48.316: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc" in namespace "projected-1565" to be "Succeeded or Failed" -Dec 14 15:59:48.324: INFO: Pod "pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc": Phase="Pending", Reason="", readiness=false. Elapsed: 7.867811ms -Dec 14 15:59:50.338: INFO: Pod "pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021839749s -Dec 14 15:59:52.336: INFO: Pod "pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020445144s -STEP: Saw pod success 12/14/22 15:59:52.337 -Dec 14 15:59:52.338: INFO: Pod "pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc" satisfied condition "Succeeded or Failed" -Dec 14 15:59:52.345: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc container projected-secret-volume-test: -STEP: delete the pod 12/14/22 15:59:52.355 -Dec 14 15:59:52.371: INFO: Waiting for pod pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc to disappear -Dec 14 15:59:52.376: INFO: Pod pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc no longer exists -[AfterEach] [sig-storage] Projected secret +STEP: Creating a kubernetes client 02/25/23 10:02:04.532 +Feb 25 10:02:04.532: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir 02/25/23 10:02:04.535 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:02:04.559 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:02:04.563 +[It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:196 +STEP: Creating a pod to test emptydir 0644 on node default medium 02/25/23 10:02:04.567 +Feb 25 10:02:04.577: INFO: Waiting up to 5m0s for pod "pod-581aedab-f084-493a-afca-2bdf037c2a63" in namespace "emptydir-9458" to be "Succeeded or Failed" +Feb 25 10:02:04.583: INFO: Pod "pod-581aedab-f084-493a-afca-2bdf037c2a63": Phase="Pending", Reason="", readiness=false. Elapsed: 5.736339ms +Feb 25 10:02:06.590: INFO: Pod "pod-581aedab-f084-493a-afca-2bdf037c2a63": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012789553s +Feb 25 10:02:08.591: INFO: Pod "pod-581aedab-f084-493a-afca-2bdf037c2a63": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013675356s +STEP: Saw pod success 02/25/23 10:02:08.591 +Feb 25 10:02:08.592: INFO: Pod "pod-581aedab-f084-493a-afca-2bdf037c2a63" satisfied condition "Succeeded or Failed" +Feb 25 10:02:08.598: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-581aedab-f084-493a-afca-2bdf037c2a63 container test-container: +STEP: delete the pod 02/25/23 10:02:08.611 +Feb 25 10:02:08.641: INFO: Waiting for pod pod-581aedab-f084-493a-afca-2bdf037c2a63 to disappear +Feb 25 10:02:08.648: INFO: Pod pod-581aedab-f084-493a-afca-2bdf037c2a63 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 -Dec 14 15:59:52.377: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-1565" for this suite. 12/14/22 15:59:52.393 -{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","completed":67,"skipped":1204,"failed":0} +Feb 25 10:02:08.649: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-9458" for this suite. 02/25/23 10:02:08.656 +{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","completed":67,"skipped":1354,"failed":0} ------------------------------ -• [4.178 seconds] -[sig-storage] Projected secret +• [4.135 seconds] +[sig-storage] EmptyDir volumes test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:77 + should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:196 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected secret + [BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:59:48.258 - Dec 14 15:59:48.258: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 15:59:48.26 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:48.283 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:48.288 - [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:77 - STEP: Creating projection with secret that has name projected-secret-test-map-f811e5ad-cd33-45cb-83ef-f136f8cd2f11 12/14/22 15:59:48.292 - STEP: Creating a pod to test consume secrets 12/14/22 15:59:48.298 - Dec 14 15:59:48.316: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc" in namespace "projected-1565" to be "Succeeded or Failed" - Dec 14 15:59:48.324: INFO: Pod "pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc": Phase="Pending", Reason="", readiness=false. Elapsed: 7.867811ms - Dec 14 15:59:50.338: INFO: Pod "pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021839749s - Dec 14 15:59:52.336: INFO: Pod "pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020445144s - STEP: Saw pod success 12/14/22 15:59:52.337 - Dec 14 15:59:52.338: INFO: Pod "pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc" satisfied condition "Succeeded or Failed" - Dec 14 15:59:52.345: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc container projected-secret-volume-test: - STEP: delete the pod 12/14/22 15:59:52.355 - Dec 14 15:59:52.371: INFO: Waiting for pod pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc to disappear - Dec 14 15:59:52.376: INFO: Pod pod-projected-secrets-827fcdb0-e229-47cc-ac8e-484328946bbc no longer exists - [AfterEach] [sig-storage] Projected secret + STEP: Creating a kubernetes client 02/25/23 10:02:04.532 + Feb 25 10:02:04.532: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir 02/25/23 10:02:04.535 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:02:04.559 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:02:04.563 + [It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:196 + STEP: Creating a pod to test emptydir 0644 on node default medium 02/25/23 10:02:04.567 + Feb 25 10:02:04.577: INFO: Waiting up to 5m0s for pod "pod-581aedab-f084-493a-afca-2bdf037c2a63" in namespace "emptydir-9458" to be "Succeeded or Failed" + Feb 25 10:02:04.583: INFO: Pod "pod-581aedab-f084-493a-afca-2bdf037c2a63": Phase="Pending", Reason="", readiness=false. Elapsed: 5.736339ms + Feb 25 10:02:06.590: INFO: Pod "pod-581aedab-f084-493a-afca-2bdf037c2a63": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012789553s + Feb 25 10:02:08.591: INFO: Pod "pod-581aedab-f084-493a-afca-2bdf037c2a63": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013675356s + STEP: Saw pod success 02/25/23 10:02:08.591 + Feb 25 10:02:08.592: INFO: Pod "pod-581aedab-f084-493a-afca-2bdf037c2a63" satisfied condition "Succeeded or Failed" + Feb 25 10:02:08.598: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-581aedab-f084-493a-afca-2bdf037c2a63 container test-container: + STEP: delete the pod 02/25/23 10:02:08.611 + Feb 25 10:02:08.641: INFO: Waiting for pod pod-581aedab-f084-493a-afca-2bdf037c2a63 to disappear + Feb 25 10:02:08.648: INFO: Pod pod-581aedab-f084-493a-afca-2bdf037c2a63 no longer exists + [AfterEach] [sig-storage] EmptyDir volumes + test/e2e/framework/framework.go:187 + Feb 25 10:02:08.649: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-9458" for this suite. 02/25/23 10:02:08.656 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-apps] Job + should delete a job [Conformance] + test/e2e/apps/job.go:309 +[BeforeEach] [sig-apps] Job + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:02:08.679 +Feb 25 10:02:08.679: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename job 02/25/23 10:02:08.681 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:02:08.708 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:02:08.714 +[It] should delete a job [Conformance] + test/e2e/apps/job.go:309 +STEP: Creating a job 02/25/23 10:02:08.72 +STEP: Ensuring active pods == parallelism 02/25/23 10:02:08.734 +STEP: delete a job 02/25/23 10:02:12.744 +STEP: deleting Job.batch foo in namespace job-4067, will wait for the garbage collector to delete the pods 02/25/23 10:02:12.744 +Feb 25 10:02:12.820: INFO: Deleting Job.batch foo took: 16.95918ms +Feb 25 10:02:12.922: INFO: Terminating Job.batch foo pods took: 101.371198ms +STEP: Ensuring job was deleted 02/25/23 10:02:44.322 +[AfterEach] [sig-apps] Job + test/e2e/framework/framework.go:187 +Feb 25 10:02:44.334: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "job-4067" for this suite. 02/25/23 10:02:44.343 +{"msg":"PASSED [sig-apps] Job should delete a job [Conformance]","completed":68,"skipped":1390,"failed":0} +------------------------------ +• [SLOW TEST] [35.678 seconds] +[sig-apps] Job +test/e2e/apps/framework.go:23 + should delete a job [Conformance] + test/e2e/apps/job.go:309 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-apps] Job + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:02:08.679 + Feb 25 10:02:08.679: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename job 02/25/23 10:02:08.681 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:02:08.708 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:02:08.714 + [It] should delete a job [Conformance] + test/e2e/apps/job.go:309 + STEP: Creating a job 02/25/23 10:02:08.72 + STEP: Ensuring active pods == parallelism 02/25/23 10:02:08.734 + STEP: delete a job 02/25/23 10:02:12.744 + STEP: deleting Job.batch foo in namespace job-4067, will wait for the garbage collector to delete the pods 02/25/23 10:02:12.744 + Feb 25 10:02:12.820: INFO: Deleting Job.batch foo took: 16.95918ms + Feb 25 10:02:12.922: INFO: Terminating Job.batch foo pods took: 101.371198ms + STEP: Ensuring job was deleted 02/25/23 10:02:44.322 + [AfterEach] [sig-apps] Job test/e2e/framework/framework.go:187 - Dec 14 15:59:52.377: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-1565" for this suite. 12/14/22 15:59:52.393 + Feb 25 10:02:44.334: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "job-4067" for this suite. 02/25/23 10:02:44.343 << End Captured GinkgoWriter Output ------------------------------ SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ +[sig-apps] ReplicationController + should test the lifecycle of a ReplicationController [Conformance] + test/e2e/apps/rc.go:109 +[BeforeEach] [sig-apps] ReplicationController + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:02:44.365 +Feb 25 10:02:44.365: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename replication-controller 02/25/23 10:02:44.368 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:02:44.411 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:02:44.416 +[BeforeEach] [sig-apps] ReplicationController + test/e2e/apps/rc.go:56 +[It] should test the lifecycle of a ReplicationController [Conformance] + test/e2e/apps/rc.go:109 +STEP: creating a ReplicationController 02/25/23 10:02:44.429 +STEP: waiting for RC to be added 02/25/23 10:02:44.443 +STEP: waiting for available Replicas 02/25/23 10:02:44.445 +STEP: patching ReplicationController 02/25/23 10:02:49.3 +STEP: waiting for RC to be modified 02/25/23 10:02:49.333 +STEP: patching ReplicationController status 02/25/23 10:02:49.333 +STEP: waiting for RC to be modified 02/25/23 10:02:49.349 +STEP: waiting for available Replicas 02/25/23 10:02:49.35 +STEP: fetching ReplicationController status 02/25/23 10:02:49.365 +STEP: patching ReplicationController scale 02/25/23 10:02:49.373 +STEP: waiting for RC to be modified 02/25/23 10:02:49.386 +STEP: waiting for ReplicationController's scale to be the max amount 02/25/23 10:02:49.387 +STEP: fetching ReplicationController; ensuring that it's patched 02/25/23 10:02:59.48 +STEP: updating ReplicationController status 02/25/23 10:02:59.49 +STEP: waiting for RC to be modified 02/25/23 10:02:59.505 +STEP: listing all ReplicationControllers 02/25/23 10:02:59.505 +STEP: checking that ReplicationController has expected values 02/25/23 10:02:59.513 +STEP: deleting ReplicationControllers by collection 02/25/23 10:02:59.514 +STEP: waiting for ReplicationController to have a DELETED watchEvent 02/25/23 10:02:59.532 +[AfterEach] [sig-apps] ReplicationController + test/e2e/framework/framework.go:187 +Feb 25 10:02:59.689: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "replication-controller-4293" for this suite. 02/25/23 10:02:59.697 +{"msg":"PASSED [sig-apps] ReplicationController should test the lifecycle of a ReplicationController [Conformance]","completed":69,"skipped":1414,"failed":0} +------------------------------ +• [SLOW TEST] [15.344 seconds] +[sig-apps] ReplicationController +test/e2e/apps/framework.go:23 + should test the lifecycle of a ReplicationController [Conformance] + test/e2e/apps/rc.go:109 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-apps] ReplicationController + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:02:44.365 + Feb 25 10:02:44.365: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename replication-controller 02/25/23 10:02:44.368 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:02:44.411 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:02:44.416 + [BeforeEach] [sig-apps] ReplicationController + test/e2e/apps/rc.go:56 + [It] should test the lifecycle of a ReplicationController [Conformance] + test/e2e/apps/rc.go:109 + STEP: creating a ReplicationController 02/25/23 10:02:44.429 + STEP: waiting for RC to be added 02/25/23 10:02:44.443 + STEP: waiting for available Replicas 02/25/23 10:02:44.445 + STEP: patching ReplicationController 02/25/23 10:02:49.3 + STEP: waiting for RC to be modified 02/25/23 10:02:49.333 + STEP: patching ReplicationController status 02/25/23 10:02:49.333 + STEP: waiting for RC to be modified 02/25/23 10:02:49.349 + STEP: waiting for available Replicas 02/25/23 10:02:49.35 + STEP: fetching ReplicationController status 02/25/23 10:02:49.365 + STEP: patching ReplicationController scale 02/25/23 10:02:49.373 + STEP: waiting for RC to be modified 02/25/23 10:02:49.386 + STEP: waiting for ReplicationController's scale to be the max amount 02/25/23 10:02:49.387 + STEP: fetching ReplicationController; ensuring that it's patched 02/25/23 10:02:59.48 + STEP: updating ReplicationController status 02/25/23 10:02:59.49 + STEP: waiting for RC to be modified 02/25/23 10:02:59.505 + STEP: listing all ReplicationControllers 02/25/23 10:02:59.505 + STEP: checking that ReplicationController has expected values 02/25/23 10:02:59.513 + STEP: deleting ReplicationControllers by collection 02/25/23 10:02:59.514 + STEP: waiting for ReplicationController to have a DELETED watchEvent 02/25/23 10:02:59.532 + [AfterEach] [sig-apps] ReplicationController + test/e2e/framework/framework.go:187 + Feb 25 10:02:59.689: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "replication-controller-4293" for this suite. 02/25/23 10:02:59.697 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSSSS +------------------------------ [sig-node] InitContainer [NodeConformance] - should invoke init containers on a RestartNever pod [Conformance] - test/e2e/common/node/init_container.go:176 + should invoke init containers on a RestartAlways pod [Conformance] + test/e2e/common/node/init_container.go:254 [BeforeEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:59:52.446 -Dec 14 15:59:52.447: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename init-container 12/14/22 15:59:52.449 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:52.476 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:52.479 +STEP: Creating a kubernetes client 02/25/23 10:02:59.714 +Feb 25 10:02:59.715: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename init-container 02/25/23 10:02:59.717 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:02:59.794 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:02:59.801 [BeforeEach] [sig-node] InitContainer [NodeConformance] test/e2e/common/node/init_container.go:164 -[It] should invoke init containers on a RestartNever pod [Conformance] - test/e2e/common/node/init_container.go:176 -STEP: creating the pod 12/14/22 15:59:52.483 -Dec 14 15:59:52.484: INFO: PodSpec: initContainers in spec.initContainers +[It] should invoke init containers on a RestartAlways pod [Conformance] + test/e2e/common/node/init_container.go:254 +STEP: creating the pod 02/25/23 10:02:59.807 +Feb 25 10:02:59.807: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:187 -Dec 14 15:59:57.337: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "init-container-8678" for this suite. 12/14/22 15:59:57.346 -{"msg":"PASSED [sig-node] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance]","completed":68,"skipped":1228,"failed":0} +Feb 25 10:03:03.336: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "init-container-2791" for this suite. 02/25/23 10:03:03.35 +{"msg":"PASSED [sig-node] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance]","completed":70,"skipped":1431,"failed":0} ------------------------------ -• [4.908 seconds] +• [3.651 seconds] [sig-node] InitContainer [NodeConformance] test/e2e/common/node/framework.go:23 - should invoke init containers on a RestartNever pod [Conformance] - test/e2e/common/node/init_container.go:176 + should invoke init containers on a RestartAlways pod [Conformance] + test/e2e/common/node/init_container.go:254 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:59:52.446 - Dec 14 15:59:52.447: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename init-container 12/14/22 15:59:52.449 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:52.476 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:52.479 + STEP: Creating a kubernetes client 02/25/23 10:02:59.714 + Feb 25 10:02:59.715: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename init-container 02/25/23 10:02:59.717 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:02:59.794 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:02:59.801 [BeforeEach] [sig-node] InitContainer [NodeConformance] test/e2e/common/node/init_container.go:164 - [It] should invoke init containers on a RestartNever pod [Conformance] - test/e2e/common/node/init_container.go:176 - STEP: creating the pod 12/14/22 15:59:52.483 - Dec 14 15:59:52.484: INFO: PodSpec: initContainers in spec.initContainers + [It] should invoke init containers on a RestartAlways pod [Conformance] + test/e2e/common/node/init_container.go:254 + STEP: creating the pod 02/25/23 10:02:59.807 + Feb 25 10:02:59.807: INFO: PodSpec: initContainers in spec.initContainers [AfterEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:187 - Dec 14 15:59:57.337: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "init-container-8678" for this suite. 12/14/22 15:59:57.346 + Feb 25 10:03:03.336: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "init-container-2791" for this suite. 02/25/23 10:03:03.35 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSS +SSSSSSSSSSSSS ------------------------------ -[sig-auth] ServiceAccounts - should allow opting out of API token automount [Conformance] - test/e2e/auth/service_accounts.go:158 -[BeforeEach] [sig-auth] ServiceAccounts +[sig-node] Probing container + should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:211 +[BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:59:57.358 -Dec 14 15:59:57.358: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename svcaccounts 12/14/22 15:59:57.36 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:57.385 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:57.388 -[It] should allow opting out of API token automount [Conformance] - test/e2e/auth/service_accounts.go:158 -Dec 14 15:59:57.418: INFO: created pod pod-service-account-defaultsa -Dec 14 15:59:57.418: INFO: pod pod-service-account-defaultsa service account token volume mount: true -Dec 14 15:59:57.431: INFO: created pod pod-service-account-mountsa -Dec 14 15:59:57.431: INFO: pod pod-service-account-mountsa service account token volume mount: true -Dec 14 15:59:57.445: INFO: created pod pod-service-account-nomountsa -Dec 14 15:59:57.445: INFO: pod pod-service-account-nomountsa service account token volume mount: false -Dec 14 15:59:57.472: INFO: created pod pod-service-account-defaultsa-mountspec -Dec 14 15:59:57.472: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true -Dec 14 15:59:57.482: INFO: created pod pod-service-account-mountsa-mountspec -Dec 14 15:59:57.482: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true -Dec 14 15:59:57.497: INFO: created pod pod-service-account-nomountsa-mountspec -Dec 14 15:59:57.497: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true -Dec 14 15:59:57.529: INFO: created pod pod-service-account-defaultsa-nomountspec -Dec 14 15:59:57.529: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false -Dec 14 15:59:57.553: INFO: created pod pod-service-account-mountsa-nomountspec -Dec 14 15:59:57.553: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false -Dec 14 15:59:57.567: INFO: created pod pod-service-account-nomountsa-nomountspec -Dec 14 15:59:57.568: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false -[AfterEach] [sig-auth] ServiceAccounts +STEP: Creating a kubernetes client 02/25/23 10:03:03.372 +Feb 25 10:03:03.373: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-probe 02/25/23 10:03:03.374 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:03:03.408 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:03:03.414 +[BeforeEach] [sig-node] Probing container + test/e2e/common/node/container_probe.go:59 +[It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:211 +STEP: Creating pod test-webserver-0e44adc6-cbf4-482a-b166-d5e3e033d179 in namespace container-probe-1545 02/25/23 10:03:03.421 +Feb 25 10:03:03.443: INFO: Waiting up to 5m0s for pod "test-webserver-0e44adc6-cbf4-482a-b166-d5e3e033d179" in namespace "container-probe-1545" to be "not pending" +Feb 25 10:03:03.463: INFO: Pod "test-webserver-0e44adc6-cbf4-482a-b166-d5e3e033d179": Phase="Pending", Reason="", readiness=false. Elapsed: 19.809107ms +Feb 25 10:03:05.481: INFO: Pod "test-webserver-0e44adc6-cbf4-482a-b166-d5e3e033d179": Phase="Running", Reason="", readiness=true. Elapsed: 2.038246007s +Feb 25 10:03:05.482: INFO: Pod "test-webserver-0e44adc6-cbf4-482a-b166-d5e3e033d179" satisfied condition "not pending" +Feb 25 10:03:05.482: INFO: Started pod test-webserver-0e44adc6-cbf4-482a-b166-d5e3e033d179 in namespace container-probe-1545 +STEP: checking the pod's current state and verifying that restartCount is present 02/25/23 10:03:05.482 +Feb 25 10:03:05.489: INFO: Initial restart count of pod test-webserver-0e44adc6-cbf4-482a-b166-d5e3e033d179 is 0 +STEP: deleting the pod 02/25/23 10:07:06.704 +[AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 -Dec 14 15:59:57.568: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "svcaccounts-9033" for this suite. 12/14/22 15:59:57.59 -{"msg":"PASSED [sig-auth] ServiceAccounts should allow opting out of API token automount [Conformance]","completed":69,"skipped":1242,"failed":0} +Feb 25 10:07:06.737: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-1545" for this suite. 02/25/23 10:07:06.756 +{"msg":"PASSED [sig-node] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","completed":71,"skipped":1444,"failed":0} ------------------------------ -• [0.261 seconds] -[sig-auth] ServiceAccounts -test/e2e/auth/framework.go:23 - should allow opting out of API token automount [Conformance] - test/e2e/auth/service_accounts.go:158 +• [SLOW TEST] [243.428 seconds] +[sig-node] Probing container +test/e2e/common/node/framework.go:23 + should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:211 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-auth] ServiceAccounts + [BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:59:57.358 - Dec 14 15:59:57.358: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename svcaccounts 12/14/22 15:59:57.36 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:57.385 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:57.388 - [It] should allow opting out of API token automount [Conformance] - test/e2e/auth/service_accounts.go:158 - Dec 14 15:59:57.418: INFO: created pod pod-service-account-defaultsa - Dec 14 15:59:57.418: INFO: pod pod-service-account-defaultsa service account token volume mount: true - Dec 14 15:59:57.431: INFO: created pod pod-service-account-mountsa - Dec 14 15:59:57.431: INFO: pod pod-service-account-mountsa service account token volume mount: true - Dec 14 15:59:57.445: INFO: created pod pod-service-account-nomountsa - Dec 14 15:59:57.445: INFO: pod pod-service-account-nomountsa service account token volume mount: false - Dec 14 15:59:57.472: INFO: created pod pod-service-account-defaultsa-mountspec - Dec 14 15:59:57.472: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true - Dec 14 15:59:57.482: INFO: created pod pod-service-account-mountsa-mountspec - Dec 14 15:59:57.482: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true - Dec 14 15:59:57.497: INFO: created pod pod-service-account-nomountsa-mountspec - Dec 14 15:59:57.497: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true - Dec 14 15:59:57.529: INFO: created pod pod-service-account-defaultsa-nomountspec - Dec 14 15:59:57.529: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false - Dec 14 15:59:57.553: INFO: created pod pod-service-account-mountsa-nomountspec - Dec 14 15:59:57.553: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false - Dec 14 15:59:57.567: INFO: created pod pod-service-account-nomountsa-nomountspec - Dec 14 15:59:57.568: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false - [AfterEach] [sig-auth] ServiceAccounts + STEP: Creating a kubernetes client 02/25/23 10:03:03.372 + Feb 25 10:03:03.373: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-probe 02/25/23 10:03:03.374 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:03:03.408 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:03:03.414 + [BeforeEach] [sig-node] Probing container + test/e2e/common/node/container_probe.go:59 + [It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:211 + STEP: Creating pod test-webserver-0e44adc6-cbf4-482a-b166-d5e3e033d179 in namespace container-probe-1545 02/25/23 10:03:03.421 + Feb 25 10:03:03.443: INFO: Waiting up to 5m0s for pod "test-webserver-0e44adc6-cbf4-482a-b166-d5e3e033d179" in namespace "container-probe-1545" to be "not pending" + Feb 25 10:03:03.463: INFO: Pod "test-webserver-0e44adc6-cbf4-482a-b166-d5e3e033d179": Phase="Pending", Reason="", readiness=false. Elapsed: 19.809107ms + Feb 25 10:03:05.481: INFO: Pod "test-webserver-0e44adc6-cbf4-482a-b166-d5e3e033d179": Phase="Running", Reason="", readiness=true. Elapsed: 2.038246007s + Feb 25 10:03:05.482: INFO: Pod "test-webserver-0e44adc6-cbf4-482a-b166-d5e3e033d179" satisfied condition "not pending" + Feb 25 10:03:05.482: INFO: Started pod test-webserver-0e44adc6-cbf4-482a-b166-d5e3e033d179 in namespace container-probe-1545 + STEP: checking the pod's current state and verifying that restartCount is present 02/25/23 10:03:05.482 + Feb 25 10:03:05.489: INFO: Initial restart count of pod test-webserver-0e44adc6-cbf4-482a-b166-d5e3e033d179 is 0 + STEP: deleting the pod 02/25/23 10:07:06.704 + [AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 - Dec 14 15:59:57.568: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "svcaccounts-9033" for this suite. 12/14/22 15:59:57.59 + Feb 25 10:07:06.737: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-probe-1545" for this suite. 02/25/23 10:07:06.756 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] PodTemplates - should delete a collection of pod templates [Conformance] - test/e2e/common/node/podtemplates.go:122 -[BeforeEach] [sig-node] PodTemplates +[sig-node] Probing container + with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:104 +[BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:59:57.621 -Dec 14 15:59:57.621: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename podtemplate 12/14/22 15:59:57.622 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:57.682 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:57.687 -[It] should delete a collection of pod templates [Conformance] - test/e2e/common/node/podtemplates.go:122 -STEP: Create set of pod templates 12/14/22 15:59:57.727 -Dec 14 15:59:57.744: INFO: created test-podtemplate-1 -Dec 14 15:59:57.758: INFO: created test-podtemplate-2 -Dec 14 15:59:57.776: INFO: created test-podtemplate-3 -STEP: get a list of pod templates with a label in the current namespace 12/14/22 15:59:57.776 -STEP: delete collection of pod templates 12/14/22 15:59:57.784 -Dec 14 15:59:57.785: INFO: requesting DeleteCollection of pod templates -STEP: check that the list of pod templates matches the requested quantity 12/14/22 15:59:57.849 -Dec 14 15:59:57.849: INFO: requesting list of pod templates to confirm quantity -[AfterEach] [sig-node] PodTemplates +STEP: Creating a kubernetes client 02/25/23 10:07:06.809 +Feb 25 10:07:06.809: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-probe 02/25/23 10:07:06.826 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:07:06.875 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:07:06.883 +[BeforeEach] [sig-node] Probing container + test/e2e/common/node/container_probe.go:59 +[It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:104 +[AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 -Dec 14 15:59:57.876: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "podtemplate-1486" for this suite. 12/14/22 15:59:57.883 -{"msg":"PASSED [sig-node] PodTemplates should delete a collection of pod templates [Conformance]","completed":70,"skipped":1277,"failed":0} +Feb 25 10:08:06.933: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-8049" for this suite. 02/25/23 10:08:06.953 +{"msg":"PASSED [sig-node] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]","completed":72,"skipped":1462,"failed":0} ------------------------------ -• [0.275 seconds] -[sig-node] PodTemplates +• [SLOW TEST] [60.178 seconds] +[sig-node] Probing container test/e2e/common/node/framework.go:23 - should delete a collection of pod templates [Conformance] - test/e2e/common/node/podtemplates.go:122 + with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:104 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] PodTemplates + [BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:59:57.621 - Dec 14 15:59:57.621: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename podtemplate 12/14/22 15:59:57.622 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:57.682 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:57.687 - [It] should delete a collection of pod templates [Conformance] - test/e2e/common/node/podtemplates.go:122 - STEP: Create set of pod templates 12/14/22 15:59:57.727 - Dec 14 15:59:57.744: INFO: created test-podtemplate-1 - Dec 14 15:59:57.758: INFO: created test-podtemplate-2 - Dec 14 15:59:57.776: INFO: created test-podtemplate-3 - STEP: get a list of pod templates with a label in the current namespace 12/14/22 15:59:57.776 - STEP: delete collection of pod templates 12/14/22 15:59:57.784 - Dec 14 15:59:57.785: INFO: requesting DeleteCollection of pod templates - STEP: check that the list of pod templates matches the requested quantity 12/14/22 15:59:57.849 - Dec 14 15:59:57.849: INFO: requesting list of pod templates to confirm quantity - [AfterEach] [sig-node] PodTemplates + STEP: Creating a kubernetes client 02/25/23 10:07:06.809 + Feb 25 10:07:06.809: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-probe 02/25/23 10:07:06.826 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:07:06.875 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:07:06.883 + [BeforeEach] [sig-node] Probing container + test/e2e/common/node/container_probe.go:59 + [It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:104 + [AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 - Dec 14 15:59:57.876: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "podtemplate-1486" for this suite. 12/14/22 15:59:57.883 + Feb 25 10:08:06.933: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-probe-8049" for this suite. 02/25/23 10:08:06.953 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] Garbage collector - should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] - test/e2e/apimachinery/garbage_collector.go:735 -[BeforeEach] [sig-api-machinery] Garbage collector +[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] + Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] + test/e2e/apps/statefulset.go:585 +[BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 15:59:57.9 -Dec 14 15:59:57.900: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename gc 12/14/22 15:59:57.903 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:57.949 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:57.953 -[It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] - test/e2e/apimachinery/garbage_collector.go:735 -STEP: create the rc1 12/14/22 15:59:57.962 -STEP: create the rc2 12/14/22 15:59:57.98 -STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well 12/14/22 16:00:03.018 -STEP: delete the rc simpletest-rc-to-be-deleted 12/14/22 16:00:04.678 -STEP: wait for the rc to be deleted 12/14/22 16:00:04.708 -Dec 14 16:00:09.737: INFO: 67 pods remaining -Dec 14 16:00:09.738: INFO: 67 pods has nil DeletionTimestamp -Dec 14 16:00:09.738: INFO: -STEP: Gathering metrics 12/14/22 16:00:14.739 -Dec 14 16:00:14.775: INFO: Waiting up to 5m0s for pod "kube-controller-manager-kiedai4iakae-2" in namespace "kube-system" to be "running and ready" -Dec 14 16:00:14.784: INFO: Pod "kube-controller-manager-kiedai4iakae-2": Phase="Running", Reason="", readiness=true. Elapsed: 8.091751ms -Dec 14 16:00:14.784: INFO: The phase of Pod kube-controller-manager-kiedai4iakae-2 is Running (Ready = true) -Dec 14 16:00:14.784: INFO: Pod "kube-controller-manager-kiedai4iakae-2" satisfied condition "running and ready" -Dec 14 16:00:15.231: INFO: For apiserver_request_total: -For apiserver_request_latency_seconds: -For apiserver_init_events_total: -For garbage_collector_attempt_to_delete_queue_latency: -For garbage_collector_attempt_to_delete_work_duration: -For garbage_collector_attempt_to_orphan_queue_latency: -For garbage_collector_attempt_to_orphan_work_duration: -For garbage_collector_dirty_processing_latency_microseconds: -For garbage_collector_event_processing_latency_microseconds: -For garbage_collector_graph_changes_queue_latency: -For garbage_collector_graph_changes_work_duration: -For garbage_collector_orphan_processing_latency_microseconds: -For namespace_queue_latency: -For namespace_queue_latency_sum: -For namespace_queue_latency_count: -For namespace_retries: -For namespace_work_duration: -For namespace_work_duration_sum: -For namespace_work_duration_count: -For function_duration_seconds: -For errors_total: -For evicted_pods_total: - -Dec 14 16:00:15.231: INFO: Deleting pod "simpletest-rc-to-be-deleted-25mtj" in namespace "gc-5605" -Dec 14 16:00:15.259: INFO: Deleting pod "simpletest-rc-to-be-deleted-26xjx" in namespace "gc-5605" -Dec 14 16:00:15.301: INFO: Deleting pod "simpletest-rc-to-be-deleted-2c5gb" in namespace "gc-5605" -Dec 14 16:00:15.341: INFO: Deleting pod "simpletest-rc-to-be-deleted-2xxxx" in namespace "gc-5605" -Dec 14 16:00:15.403: INFO: Deleting pod "simpletest-rc-to-be-deleted-546gb" in namespace "gc-5605" -Dec 14 16:00:15.458: INFO: Deleting pod "simpletest-rc-to-be-deleted-5djvj" in namespace "gc-5605" -Dec 14 16:00:15.552: INFO: Deleting pod "simpletest-rc-to-be-deleted-5h6v4" in namespace "gc-5605" -Dec 14 16:00:15.656: INFO: Deleting pod "simpletest-rc-to-be-deleted-5pv72" in namespace "gc-5605" -Dec 14 16:00:15.686: INFO: Deleting pod "simpletest-rc-to-be-deleted-5xr2d" in namespace "gc-5605" -Dec 14 16:00:15.719: INFO: Deleting pod "simpletest-rc-to-be-deleted-65495" in namespace "gc-5605" -Dec 14 16:00:15.783: INFO: Deleting pod "simpletest-rc-to-be-deleted-69kd6" in namespace "gc-5605" -Dec 14 16:00:15.827: INFO: Deleting pod "simpletest-rc-to-be-deleted-6j9zj" in namespace "gc-5605" -Dec 14 16:00:15.908: INFO: Deleting pod "simpletest-rc-to-be-deleted-6tqsv" in namespace "gc-5605" -Dec 14 16:00:15.936: INFO: Deleting pod "simpletest-rc-to-be-deleted-74pvv" in namespace "gc-5605" -Dec 14 16:00:15.954: INFO: Deleting pod "simpletest-rc-to-be-deleted-7hc59" in namespace "gc-5605" -Dec 14 16:00:15.977: INFO: Deleting pod "simpletest-rc-to-be-deleted-7pdx2" in namespace "gc-5605" -Dec 14 16:00:16.007: INFO: Deleting pod "simpletest-rc-to-be-deleted-7rgbh" in namespace "gc-5605" -Dec 14 16:00:16.056: INFO: Deleting pod "simpletest-rc-to-be-deleted-88tjd" in namespace "gc-5605" -Dec 14 16:00:16.133: INFO: Deleting pod "simpletest-rc-to-be-deleted-8xzz4" in namespace "gc-5605" -Dec 14 16:00:16.176: INFO: Deleting pod "simpletest-rc-to-be-deleted-92tfn" in namespace "gc-5605" -Dec 14 16:00:16.287: INFO: Deleting pod "simpletest-rc-to-be-deleted-97mr2" in namespace "gc-5605" -Dec 14 16:00:16.357: INFO: Deleting pod "simpletest-rc-to-be-deleted-98dmm" in namespace "gc-5605" -Dec 14 16:00:16.400: INFO: Deleting pod "simpletest-rc-to-be-deleted-9d26n" in namespace "gc-5605" -Dec 14 16:00:16.418: INFO: Deleting pod "simpletest-rc-to-be-deleted-9np8l" in namespace "gc-5605" -Dec 14 16:00:16.462: INFO: Deleting pod "simpletest-rc-to-be-deleted-9wzjl" in namespace "gc-5605" -Dec 14 16:00:16.512: INFO: Deleting pod "simpletest-rc-to-be-deleted-9xc5j" in namespace "gc-5605" -Dec 14 16:00:16.570: INFO: Deleting pod "simpletest-rc-to-be-deleted-9zktc" in namespace "gc-5605" -Dec 14 16:00:16.610: INFO: Deleting pod "simpletest-rc-to-be-deleted-b6kmj" in namespace "gc-5605" -Dec 14 16:00:16.644: INFO: Deleting pod "simpletest-rc-to-be-deleted-b8jrk" in namespace "gc-5605" -Dec 14 16:00:16.687: INFO: Deleting pod "simpletest-rc-to-be-deleted-bnssz" in namespace "gc-5605" -Dec 14 16:00:16.759: INFO: Deleting pod "simpletest-rc-to-be-deleted-bpq8l" in namespace "gc-5605" -Dec 14 16:00:16.806: INFO: Deleting pod "simpletest-rc-to-be-deleted-bwxft" in namespace "gc-5605" -Dec 14 16:00:16.833: INFO: Deleting pod "simpletest-rc-to-be-deleted-c8pck" in namespace "gc-5605" -Dec 14 16:00:16.861: INFO: Deleting pod "simpletest-rc-to-be-deleted-c9tvf" in namespace "gc-5605" -Dec 14 16:00:16.895: INFO: Deleting pod "simpletest-rc-to-be-deleted-ccqgr" in namespace "gc-5605" -Dec 14 16:00:16.950: INFO: Deleting pod "simpletest-rc-to-be-deleted-chs7g" in namespace "gc-5605" -Dec 14 16:00:17.011: INFO: Deleting pod "simpletest-rc-to-be-deleted-d22c8" in namespace "gc-5605" -Dec 14 16:00:17.044: INFO: Deleting pod "simpletest-rc-to-be-deleted-d28t4" in namespace "gc-5605" -Dec 14 16:00:17.259: INFO: Deleting pod "simpletest-rc-to-be-deleted-d7rd4" in namespace "gc-5605" -Dec 14 16:00:17.333: INFO: Deleting pod "simpletest-rc-to-be-deleted-dp9m5" in namespace "gc-5605" -Dec 14 16:00:17.395: INFO: Deleting pod "simpletest-rc-to-be-deleted-fcnk8" in namespace "gc-5605" -Dec 14 16:00:17.442: INFO: Deleting pod "simpletest-rc-to-be-deleted-fms5q" in namespace "gc-5605" -Dec 14 16:00:17.503: INFO: Deleting pod "simpletest-rc-to-be-deleted-fnwl9" in namespace "gc-5605" -Dec 14 16:00:17.553: INFO: Deleting pod "simpletest-rc-to-be-deleted-fp8v7" in namespace "gc-5605" -Dec 14 16:00:17.603: INFO: Deleting pod "simpletest-rc-to-be-deleted-g7knr" in namespace "gc-5605" -Dec 14 16:00:17.639: INFO: Deleting pod "simpletest-rc-to-be-deleted-g8dkr" in namespace "gc-5605" -Dec 14 16:00:17.669: INFO: Deleting pod "simpletest-rc-to-be-deleted-g9tj4" in namespace "gc-5605" -Dec 14 16:00:17.707: INFO: Deleting pod "simpletest-rc-to-be-deleted-gc5v2" in namespace "gc-5605" -Dec 14 16:00:17.759: INFO: Deleting pod "simpletest-rc-to-be-deleted-gdkhd" in namespace "gc-5605" -Dec 14 16:00:17.786: INFO: Deleting pod "simpletest-rc-to-be-deleted-hkfn5" in namespace "gc-5605" -[AfterEach] [sig-api-machinery] Garbage collector +STEP: Creating a kubernetes client 02/25/23 10:08:06.996 +Feb 25 10:08:06.996: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename statefulset 02/25/23 10:08:07.003 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:08:07.086 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:08:07.092 +[BeforeEach] [sig-apps] StatefulSet + test/e2e/apps/statefulset.go:96 +[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:111 +STEP: Creating service test in namespace statefulset-8614 02/25/23 10:08:07.099 +[It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] + test/e2e/apps/statefulset.go:585 +STEP: Initializing watcher for selector baz=blah,foo=bar 02/25/23 10:08:07.114 +STEP: Creating stateful set ss in namespace statefulset-8614 02/25/23 10:08:07.126 +STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-8614 02/25/23 10:08:07.141 +Feb 25 10:08:07.159: INFO: Found 0 stateful pods, waiting for 1 +Feb 25 10:08:17.177: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true +STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod 02/25/23 10:08:17.177 +Feb 25 10:08:17.190: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Feb 25 10:08:17.561: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Feb 25 10:08:17.561: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Feb 25 10:08:17.561: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +Feb 25 10:08:17.569: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true +Feb 25 10:08:27.579: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false +Feb 25 10:08:27.579: INFO: Waiting for statefulset status.replicas updated to 0 +Feb 25 10:08:27.635: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999999246s +Feb 25 10:08:28.648: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.981606105s +Feb 25 10:08:29.665: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.968393187s +Feb 25 10:08:30.676: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.952070112s +Feb 25 10:08:31.687: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.941064853s +Feb 25 10:08:32.699: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.930390051s +Feb 25 10:08:33.710: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.91766954s +Feb 25 10:08:34.719: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.907181667s +Feb 25 10:08:35.728: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.898075776s +Feb 25 10:08:36.736: INFO: Verifying statefulset ss doesn't scale past 1 for another 889.41201ms +STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-8614 02/25/23 10:08:37.737 +Feb 25 10:08:37.760: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' +Feb 25 10:08:38.273: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" +Feb 25 10:08:38.273: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" +Feb 25 10:08:38.273: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + +Feb 25 10:08:38.281: INFO: Found 1 stateful pods, waiting for 3 +Feb 25 10:08:48.294: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true +Feb 25 10:08:48.294: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true +Feb 25 10:08:48.294: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true +STEP: Verifying that stateful set ss was scaled up in order 02/25/23 10:08:48.294 +STEP: Scale down will halt with unhealthy stateful pod 02/25/23 10:08:48.295 +Feb 25 10:08:48.318: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Feb 25 10:08:48.634: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Feb 25 10:08:48.634: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Feb 25 10:08:48.634: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +Feb 25 10:08:48.635: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Feb 25 10:08:49.117: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Feb 25 10:08:49.117: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Feb 25 10:08:49.117: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +Feb 25 10:08:49.117: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Feb 25 10:08:49.683: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Feb 25 10:08:49.683: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Feb 25 10:08:49.683: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +Feb 25 10:08:49.683: INFO: Waiting for statefulset status.replicas updated to 0 +Feb 25 10:08:49.690: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 2 +Feb 25 10:08:59.715: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false +Feb 25 10:08:59.715: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false +Feb 25 10:08:59.715: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false +Feb 25 10:08:59.752: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999999182s +Feb 25 10:09:00.764: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.989042463s +Feb 25 10:09:01.772: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.977610098s +Feb 25 10:09:02.781: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.969840935s +Feb 25 10:09:03.791: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.960245416s +Feb 25 10:09:04.799: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.95039868s +Feb 25 10:09:05.811: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.942365019s +Feb 25 10:09:06.825: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.930200142s +Feb 25 10:09:07.838: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.914737154s +Feb 25 10:09:08.906: INFO: Verifying statefulset ss doesn't scale past 3 for another 902.78823ms +STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-8614 02/25/23 10:09:09.908 +Feb 25 10:09:09.927: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' +Feb 25 10:09:10.348: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" +Feb 25 10:09:10.349: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" +Feb 25 10:09:10.349: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + +Feb 25 10:09:10.349: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' +Feb 25 10:09:10.707: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" +Feb 25 10:09:10.707: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" +Feb 25 10:09:10.707: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + +Feb 25 10:09:10.707: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' +Feb 25 10:09:11.218: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" +Feb 25 10:09:11.218: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" +Feb 25 10:09:11.218: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + +Feb 25 10:09:11.218: INFO: Scaling statefulset ss to 0 +STEP: Verifying that stateful set ss was scaled down in reverse order 02/25/23 10:09:21.256 +[AfterEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:122 +Feb 25 10:09:21.257: INFO: Deleting all statefulset in ns statefulset-8614 +Feb 25 10:09:21.264: INFO: Scaling statefulset ss to 0 +Feb 25 10:09:21.289: INFO: Waiting for statefulset status.replicas updated to 0 +Feb 25 10:09:21.295: INFO: Deleting statefulset ss +[AfterEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:187 -Dec 14 16:00:17.858: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "gc-5605" for this suite. 12/14/22 16:00:17.874 -{"msg":"PASSED [sig-api-machinery] Garbage collector should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]","completed":71,"skipped":1280,"failed":0} +Feb 25 10:09:21.335: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "statefulset-8614" for this suite. 02/25/23 10:09:21.345 +{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance]","completed":73,"skipped":1478,"failed":0} ------------------------------ -• [SLOW TEST] [19.986 seconds] -[sig-api-machinery] Garbage collector -test/e2e/apimachinery/framework.go:23 - should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] - test/e2e/apimachinery/garbage_collector.go:735 +• [SLOW TEST] [74.364 seconds] +[sig-apps] StatefulSet +test/e2e/apps/framework.go:23 + Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:101 + Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] + test/e2e/apps/statefulset.go:585 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Garbage collector + [BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 15:59:57.9 - Dec 14 15:59:57.900: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename gc 12/14/22 15:59:57.903 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 15:59:57.949 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 15:59:57.953 - [It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] - test/e2e/apimachinery/garbage_collector.go:735 - STEP: create the rc1 12/14/22 15:59:57.962 - STEP: create the rc2 12/14/22 15:59:57.98 - STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well 12/14/22 16:00:03.018 - STEP: delete the rc simpletest-rc-to-be-deleted 12/14/22 16:00:04.678 - STEP: wait for the rc to be deleted 12/14/22 16:00:04.708 - Dec 14 16:00:09.737: INFO: 67 pods remaining - Dec 14 16:00:09.738: INFO: 67 pods has nil DeletionTimestamp - Dec 14 16:00:09.738: INFO: - STEP: Gathering metrics 12/14/22 16:00:14.739 - Dec 14 16:00:14.775: INFO: Waiting up to 5m0s for pod "kube-controller-manager-kiedai4iakae-2" in namespace "kube-system" to be "running and ready" - Dec 14 16:00:14.784: INFO: Pod "kube-controller-manager-kiedai4iakae-2": Phase="Running", Reason="", readiness=true. Elapsed: 8.091751ms - Dec 14 16:00:14.784: INFO: The phase of Pod kube-controller-manager-kiedai4iakae-2 is Running (Ready = true) - Dec 14 16:00:14.784: INFO: Pod "kube-controller-manager-kiedai4iakae-2" satisfied condition "running and ready" - Dec 14 16:00:15.231: INFO: For apiserver_request_total: - For apiserver_request_latency_seconds: - For apiserver_init_events_total: - For garbage_collector_attempt_to_delete_queue_latency: - For garbage_collector_attempt_to_delete_work_duration: - For garbage_collector_attempt_to_orphan_queue_latency: - For garbage_collector_attempt_to_orphan_work_duration: - For garbage_collector_dirty_processing_latency_microseconds: - For garbage_collector_event_processing_latency_microseconds: - For garbage_collector_graph_changes_queue_latency: - For garbage_collector_graph_changes_work_duration: - For garbage_collector_orphan_processing_latency_microseconds: - For namespace_queue_latency: - For namespace_queue_latency_sum: - For namespace_queue_latency_count: - For namespace_retries: - For namespace_work_duration: - For namespace_work_duration_sum: - For namespace_work_duration_count: - For function_duration_seconds: - For errors_total: - For evicted_pods_total: - - Dec 14 16:00:15.231: INFO: Deleting pod "simpletest-rc-to-be-deleted-25mtj" in namespace "gc-5605" - Dec 14 16:00:15.259: INFO: Deleting pod "simpletest-rc-to-be-deleted-26xjx" in namespace "gc-5605" - Dec 14 16:00:15.301: INFO: Deleting pod "simpletest-rc-to-be-deleted-2c5gb" in namespace "gc-5605" - Dec 14 16:00:15.341: INFO: Deleting pod "simpletest-rc-to-be-deleted-2xxxx" in namespace "gc-5605" - Dec 14 16:00:15.403: INFO: Deleting pod "simpletest-rc-to-be-deleted-546gb" in namespace "gc-5605" - Dec 14 16:00:15.458: INFO: Deleting pod "simpletest-rc-to-be-deleted-5djvj" in namespace "gc-5605" - Dec 14 16:00:15.552: INFO: Deleting pod "simpletest-rc-to-be-deleted-5h6v4" in namespace "gc-5605" - Dec 14 16:00:15.656: INFO: Deleting pod "simpletest-rc-to-be-deleted-5pv72" in namespace "gc-5605" - Dec 14 16:00:15.686: INFO: Deleting pod "simpletest-rc-to-be-deleted-5xr2d" in namespace "gc-5605" - Dec 14 16:00:15.719: INFO: Deleting pod "simpletest-rc-to-be-deleted-65495" in namespace "gc-5605" - Dec 14 16:00:15.783: INFO: Deleting pod "simpletest-rc-to-be-deleted-69kd6" in namespace "gc-5605" - Dec 14 16:00:15.827: INFO: Deleting pod "simpletest-rc-to-be-deleted-6j9zj" in namespace "gc-5605" - Dec 14 16:00:15.908: INFO: Deleting pod "simpletest-rc-to-be-deleted-6tqsv" in namespace "gc-5605" - Dec 14 16:00:15.936: INFO: Deleting pod "simpletest-rc-to-be-deleted-74pvv" in namespace "gc-5605" - Dec 14 16:00:15.954: INFO: Deleting pod "simpletest-rc-to-be-deleted-7hc59" in namespace "gc-5605" - Dec 14 16:00:15.977: INFO: Deleting pod "simpletest-rc-to-be-deleted-7pdx2" in namespace "gc-5605" - Dec 14 16:00:16.007: INFO: Deleting pod "simpletest-rc-to-be-deleted-7rgbh" in namespace "gc-5605" - Dec 14 16:00:16.056: INFO: Deleting pod "simpletest-rc-to-be-deleted-88tjd" in namespace "gc-5605" - Dec 14 16:00:16.133: INFO: Deleting pod "simpletest-rc-to-be-deleted-8xzz4" in namespace "gc-5605" - Dec 14 16:00:16.176: INFO: Deleting pod "simpletest-rc-to-be-deleted-92tfn" in namespace "gc-5605" - Dec 14 16:00:16.287: INFO: Deleting pod "simpletest-rc-to-be-deleted-97mr2" in namespace "gc-5605" - Dec 14 16:00:16.357: INFO: Deleting pod "simpletest-rc-to-be-deleted-98dmm" in namespace "gc-5605" - Dec 14 16:00:16.400: INFO: Deleting pod "simpletest-rc-to-be-deleted-9d26n" in namespace "gc-5605" - Dec 14 16:00:16.418: INFO: Deleting pod "simpletest-rc-to-be-deleted-9np8l" in namespace "gc-5605" - Dec 14 16:00:16.462: INFO: Deleting pod "simpletest-rc-to-be-deleted-9wzjl" in namespace "gc-5605" - Dec 14 16:00:16.512: INFO: Deleting pod "simpletest-rc-to-be-deleted-9xc5j" in namespace "gc-5605" - Dec 14 16:00:16.570: INFO: Deleting pod "simpletest-rc-to-be-deleted-9zktc" in namespace "gc-5605" - Dec 14 16:00:16.610: INFO: Deleting pod "simpletest-rc-to-be-deleted-b6kmj" in namespace "gc-5605" - Dec 14 16:00:16.644: INFO: Deleting pod "simpletest-rc-to-be-deleted-b8jrk" in namespace "gc-5605" - Dec 14 16:00:16.687: INFO: Deleting pod "simpletest-rc-to-be-deleted-bnssz" in namespace "gc-5605" - Dec 14 16:00:16.759: INFO: Deleting pod "simpletest-rc-to-be-deleted-bpq8l" in namespace "gc-5605" - Dec 14 16:00:16.806: INFO: Deleting pod "simpletest-rc-to-be-deleted-bwxft" in namespace "gc-5605" - Dec 14 16:00:16.833: INFO: Deleting pod "simpletest-rc-to-be-deleted-c8pck" in namespace "gc-5605" - Dec 14 16:00:16.861: INFO: Deleting pod "simpletest-rc-to-be-deleted-c9tvf" in namespace "gc-5605" - Dec 14 16:00:16.895: INFO: Deleting pod "simpletest-rc-to-be-deleted-ccqgr" in namespace "gc-5605" - Dec 14 16:00:16.950: INFO: Deleting pod "simpletest-rc-to-be-deleted-chs7g" in namespace "gc-5605" - Dec 14 16:00:17.011: INFO: Deleting pod "simpletest-rc-to-be-deleted-d22c8" in namespace "gc-5605" - Dec 14 16:00:17.044: INFO: Deleting pod "simpletest-rc-to-be-deleted-d28t4" in namespace "gc-5605" - Dec 14 16:00:17.259: INFO: Deleting pod "simpletest-rc-to-be-deleted-d7rd4" in namespace "gc-5605" - Dec 14 16:00:17.333: INFO: Deleting pod "simpletest-rc-to-be-deleted-dp9m5" in namespace "gc-5605" - Dec 14 16:00:17.395: INFO: Deleting pod "simpletest-rc-to-be-deleted-fcnk8" in namespace "gc-5605" - Dec 14 16:00:17.442: INFO: Deleting pod "simpletest-rc-to-be-deleted-fms5q" in namespace "gc-5605" - Dec 14 16:00:17.503: INFO: Deleting pod "simpletest-rc-to-be-deleted-fnwl9" in namespace "gc-5605" - Dec 14 16:00:17.553: INFO: Deleting pod "simpletest-rc-to-be-deleted-fp8v7" in namespace "gc-5605" - Dec 14 16:00:17.603: INFO: Deleting pod "simpletest-rc-to-be-deleted-g7knr" in namespace "gc-5605" - Dec 14 16:00:17.639: INFO: Deleting pod "simpletest-rc-to-be-deleted-g8dkr" in namespace "gc-5605" - Dec 14 16:00:17.669: INFO: Deleting pod "simpletest-rc-to-be-deleted-g9tj4" in namespace "gc-5605" - Dec 14 16:00:17.707: INFO: Deleting pod "simpletest-rc-to-be-deleted-gc5v2" in namespace "gc-5605" - Dec 14 16:00:17.759: INFO: Deleting pod "simpletest-rc-to-be-deleted-gdkhd" in namespace "gc-5605" - Dec 14 16:00:17.786: INFO: Deleting pod "simpletest-rc-to-be-deleted-hkfn5" in namespace "gc-5605" - [AfterEach] [sig-api-machinery] Garbage collector + STEP: Creating a kubernetes client 02/25/23 10:08:06.996 + Feb 25 10:08:06.996: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename statefulset 02/25/23 10:08:07.003 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:08:07.086 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:08:07.092 + [BeforeEach] [sig-apps] StatefulSet + test/e2e/apps/statefulset.go:96 + [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:111 + STEP: Creating service test in namespace statefulset-8614 02/25/23 10:08:07.099 + [It] Scaling should happen in predictable order and halt if any stateful pod is unhealthy [Slow] [Conformance] + test/e2e/apps/statefulset.go:585 + STEP: Initializing watcher for selector baz=blah,foo=bar 02/25/23 10:08:07.114 + STEP: Creating stateful set ss in namespace statefulset-8614 02/25/23 10:08:07.126 + STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-8614 02/25/23 10:08:07.141 + Feb 25 10:08:07.159: INFO: Found 0 stateful pods, waiting for 1 + Feb 25 10:08:17.177: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true + STEP: Confirming that stateful set scale up will halt with unhealthy stateful pod 02/25/23 10:08:17.177 + Feb 25 10:08:17.190: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' + Feb 25 10:08:17.561: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" + Feb 25 10:08:17.561: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" + Feb 25 10:08:17.561: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + + Feb 25 10:08:17.569: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true + Feb 25 10:08:27.579: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false + Feb 25 10:08:27.579: INFO: Waiting for statefulset status.replicas updated to 0 + Feb 25 10:08:27.635: INFO: Verifying statefulset ss doesn't scale past 1 for another 9.999999246s + Feb 25 10:08:28.648: INFO: Verifying statefulset ss doesn't scale past 1 for another 8.981606105s + Feb 25 10:08:29.665: INFO: Verifying statefulset ss doesn't scale past 1 for another 7.968393187s + Feb 25 10:08:30.676: INFO: Verifying statefulset ss doesn't scale past 1 for another 6.952070112s + Feb 25 10:08:31.687: INFO: Verifying statefulset ss doesn't scale past 1 for another 5.941064853s + Feb 25 10:08:32.699: INFO: Verifying statefulset ss doesn't scale past 1 for another 4.930390051s + Feb 25 10:08:33.710: INFO: Verifying statefulset ss doesn't scale past 1 for another 3.91766954s + Feb 25 10:08:34.719: INFO: Verifying statefulset ss doesn't scale past 1 for another 2.907181667s + Feb 25 10:08:35.728: INFO: Verifying statefulset ss doesn't scale past 1 for another 1.898075776s + Feb 25 10:08:36.736: INFO: Verifying statefulset ss doesn't scale past 1 for another 889.41201ms + STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-8614 02/25/23 10:08:37.737 + Feb 25 10:08:37.760: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' + Feb 25 10:08:38.273: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" + Feb 25 10:08:38.273: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" + Feb 25 10:08:38.273: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + + Feb 25 10:08:38.281: INFO: Found 1 stateful pods, waiting for 3 + Feb 25 10:08:48.294: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true + Feb 25 10:08:48.294: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true + Feb 25 10:08:48.294: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true + STEP: Verifying that stateful set ss was scaled up in order 02/25/23 10:08:48.294 + STEP: Scale down will halt with unhealthy stateful pod 02/25/23 10:08:48.295 + Feb 25 10:08:48.318: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' + Feb 25 10:08:48.634: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" + Feb 25 10:08:48.634: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" + Feb 25 10:08:48.634: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + + Feb 25 10:08:48.635: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' + Feb 25 10:08:49.117: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" + Feb 25 10:08:49.117: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" + Feb 25 10:08:49.117: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + + Feb 25 10:08:49.117: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' + Feb 25 10:08:49.683: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" + Feb 25 10:08:49.683: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" + Feb 25 10:08:49.683: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + + Feb 25 10:08:49.683: INFO: Waiting for statefulset status.replicas updated to 0 + Feb 25 10:08:49.690: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 2 + Feb 25 10:08:59.715: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false + Feb 25 10:08:59.715: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false + Feb 25 10:08:59.715: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false + Feb 25 10:08:59.752: INFO: Verifying statefulset ss doesn't scale past 3 for another 9.999999182s + Feb 25 10:09:00.764: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.989042463s + Feb 25 10:09:01.772: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.977610098s + Feb 25 10:09:02.781: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.969840935s + Feb 25 10:09:03.791: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.960245416s + Feb 25 10:09:04.799: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.95039868s + Feb 25 10:09:05.811: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.942365019s + Feb 25 10:09:06.825: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.930200142s + Feb 25 10:09:07.838: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.914737154s + Feb 25 10:09:08.906: INFO: Verifying statefulset ss doesn't scale past 3 for another 902.78823ms + STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-8614 02/25/23 10:09:09.908 + Feb 25 10:09:09.927: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' + Feb 25 10:09:10.348: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" + Feb 25 10:09:10.349: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" + Feb 25 10:09:10.349: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + + Feb 25 10:09:10.349: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' + Feb 25 10:09:10.707: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" + Feb 25 10:09:10.707: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" + Feb 25 10:09:10.707: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + + Feb 25 10:09:10.707: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-8614 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' + Feb 25 10:09:11.218: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" + Feb 25 10:09:11.218: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" + Feb 25 10:09:11.218: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + + Feb 25 10:09:11.218: INFO: Scaling statefulset ss to 0 + STEP: Verifying that stateful set ss was scaled down in reverse order 02/25/23 10:09:21.256 + [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:122 + Feb 25 10:09:21.257: INFO: Deleting all statefulset in ns statefulset-8614 + Feb 25 10:09:21.264: INFO: Scaling statefulset ss to 0 + Feb 25 10:09:21.289: INFO: Waiting for statefulset status.replicas updated to 0 + Feb 25 10:09:21.295: INFO: Deleting statefulset ss + [AfterEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:187 - Dec 14 16:00:17.858: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "gc-5605" for this suite. 12/14/22 16:00:17.874 + Feb 25 10:09:21.335: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "statefulset-8614" for this suite. 02/25/23 10:09:21.345 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSS ------------------------------ -[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition - getting/updating/patching custom resource definition status sub-resource works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:145 -[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +[sig-storage] Projected configMap + updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:123 +[BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:00:17.895 -Dec 14 16:00:17.895: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename custom-resource-definition 12/14/22 16:00:17.899 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:00:17.942 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:00:17.947 -[It] getting/updating/patching custom resource definition status sub-resource works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:145 -Dec 14 16:00:17.954: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 10:09:21.365 +Feb 25 10:09:21.365: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 10:09:21.372 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:09:21.403 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:09:21.411 +[It] updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:123 +STEP: Creating projection with configMap that has name projected-configmap-test-upd-3ea93479-f8bf-4c86-9942-fd7afe378ad9 02/25/23 10:09:21.426 +STEP: Creating the pod 02/25/23 10:09:21.436 +Feb 25 10:09:21.453: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-3476765d-d2b9-4e02-938b-307c74961a1b" in namespace "projected-4472" to be "running and ready" +Feb 25 10:09:21.461: INFO: Pod "pod-projected-configmaps-3476765d-d2b9-4e02-938b-307c74961a1b": Phase="Pending", Reason="", readiness=false. Elapsed: 7.28964ms +Feb 25 10:09:21.461: INFO: The phase of Pod pod-projected-configmaps-3476765d-d2b9-4e02-938b-307c74961a1b is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:09:23.469: INFO: Pod "pod-projected-configmaps-3476765d-d2b9-4e02-938b-307c74961a1b": Phase="Running", Reason="", readiness=true. Elapsed: 2.015313633s +Feb 25 10:09:23.469: INFO: The phase of Pod pod-projected-configmaps-3476765d-d2b9-4e02-938b-307c74961a1b is Running (Ready = true) +Feb 25 10:09:23.469: INFO: Pod "pod-projected-configmaps-3476765d-d2b9-4e02-938b-307c74961a1b" satisfied condition "running and ready" +STEP: Updating configmap projected-configmap-test-upd-3ea93479-f8bf-4c86-9942-fd7afe378ad9 02/25/23 10:09:23.515 +STEP: waiting to observe update in volume 02/25/23 10:09:23.524 +[AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 -Dec 14 16:00:18.571: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "custom-resource-definition-8537" for this suite. 12/14/22 16:00:18.589 -{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition getting/updating/patching custom resource definition status sub-resource works [Conformance]","completed":72,"skipped":1308,"failed":0} +Feb 25 10:09:25.567: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-4472" for this suite. 02/25/23 10:09:25.579 +{"msg":"PASSED [sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance]","completed":74,"skipped":1482,"failed":0} ------------------------------ -• [0.730 seconds] -[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - Simple CustomResourceDefinition - test/e2e/apimachinery/custom_resource_definition.go:50 - getting/updating/patching custom resource definition status sub-resource works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:145 +• [4.229 seconds] +[sig-storage] Projected configMap +test/e2e/common/storage/framework.go:23 + updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:123 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] + [BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:00:17.895 - Dec 14 16:00:17.895: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename custom-resource-definition 12/14/22 16:00:17.899 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:00:17.942 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:00:17.947 - [It] getting/updating/patching custom resource definition status sub-resource works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:145 - Dec 14 16:00:17.954: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 10:09:21.365 + Feb 25 10:09:21.365: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 10:09:21.372 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:09:21.403 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:09:21.411 + [It] updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:123 + STEP: Creating projection with configMap that has name projected-configmap-test-upd-3ea93479-f8bf-4c86-9942-fd7afe378ad9 02/25/23 10:09:21.426 + STEP: Creating the pod 02/25/23 10:09:21.436 + Feb 25 10:09:21.453: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-3476765d-d2b9-4e02-938b-307c74961a1b" in namespace "projected-4472" to be "running and ready" + Feb 25 10:09:21.461: INFO: Pod "pod-projected-configmaps-3476765d-d2b9-4e02-938b-307c74961a1b": Phase="Pending", Reason="", readiness=false. Elapsed: 7.28964ms + Feb 25 10:09:21.461: INFO: The phase of Pod pod-projected-configmaps-3476765d-d2b9-4e02-938b-307c74961a1b is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:09:23.469: INFO: Pod "pod-projected-configmaps-3476765d-d2b9-4e02-938b-307c74961a1b": Phase="Running", Reason="", readiness=true. Elapsed: 2.015313633s + Feb 25 10:09:23.469: INFO: The phase of Pod pod-projected-configmaps-3476765d-d2b9-4e02-938b-307c74961a1b is Running (Ready = true) + Feb 25 10:09:23.469: INFO: Pod "pod-projected-configmaps-3476765d-d2b9-4e02-938b-307c74961a1b" satisfied condition "running and ready" + STEP: Updating configmap projected-configmap-test-upd-3ea93479-f8bf-4c86-9942-fd7afe378ad9 02/25/23 10:09:23.515 + STEP: waiting to observe update in volume 02/25/23 10:09:23.524 + [AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 - Dec 14 16:00:18.571: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "custom-resource-definition-8537" for this suite. 12/14/22 16:00:18.589 + Feb 25 10:09:25.567: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-4472" for this suite. 02/25/23 10:09:25.579 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] InitContainer [NodeConformance] - should not start app containers if init containers fail on a RestartAlways pod [Conformance] - test/e2e/common/node/init_container.go:333 -[BeforeEach] [sig-node] InitContainer [NodeConformance] +[sig-storage] Projected configMap + should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:108 +[BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:00:18.64 -Dec 14 16:00:18.640: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename init-container 12/14/22 16:00:18.648 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:00:18.731 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:00:18.735 -[BeforeEach] [sig-node] InitContainer [NodeConformance] - test/e2e/common/node/init_container.go:164 -[It] should not start app containers if init containers fail on a RestartAlways pod [Conformance] - test/e2e/common/node/init_container.go:333 -STEP: creating the pod 12/14/22 16:00:18.738 -Dec 14 16:00:18.739: INFO: PodSpec: initContainers in spec.initContainers -Dec 14 16:01:07.089: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-d2447cfc-008e-493a-81ec-e393d6b36751", GenerateName:"", Namespace:"init-container-3725", SelfLink:"", UID:"861d9d60-b513-4d48-b4c8-379788bd9ccd", ResourceVersion:"11774", Generation:0, CreationTimestamp:time.Date(2022, time.December, 14, 16, 0, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"739103990"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:"e2e.test", Operation:"Update", APIVersion:"v1", Time:time.Date(2022, time.December, 14, 16, 0, 18, 0, time.Local), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc002e8cd80), Subresource:""}, v1.ManagedFieldsEntry{Manager:"kubelet", Operation:"Update", APIVersion:"v1", Time:time.Date(2022, time.December, 14, 16, 1, 7, 0, time.Local), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc002e8cdb0), Subresource:"status"}}}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"kube-api-access-gxpl2", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(nil), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(0xc00618b020), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil), Ephemeral:(*v1.EphemeralVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-api-access-gxpl2", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-api-access-gxpl2", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"registry.k8s.io/pause:3.8", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-api-access-gxpl2", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, EphemeralContainers:[]v1.EphemeralContainer(nil), RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc0040fb8d8), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"kiedai4iakae-2", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc006192e00), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0040fb970)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0040fb990)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc0040fb998), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc0040fb99c), PreemptionPolicy:(*v1.PreemptionPolicy)(0xc0039dc920), Overhead:v1.ResourceList(nil), TopologySpreadConstraints:[]v1.TopologySpreadConstraint(nil), SetHostnameAsFQDN:(*bool)(nil), OS:(*v1.PodOS)(nil), HostUsers:(*bool)(nil)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2022, time.December, 14, 16, 0, 18, 0, time.Local), Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2022, time.December, 14, 16, 0, 18, 0, time.Local), Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2022, time.December, 14, 16, 0, 18, 0, time.Local), Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2022, time.December, 14, 16, 0, 18, 0, time.Local), Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"192.168.121.197", PodIP:"10.233.65.85", PodIPs:[]v1.PodIP{v1.PodIP{IP:"10.233.65.85"}}, StartTime:time.Date(2022, time.December, 14, 16, 0, 18, 0, time.Local), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc006192ee0)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc006192f50)}, Ready:false, RestartCount:3, Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", ImageID:"registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf", ContainerID:"cri-o://6b0d606803bf8a14fb2ce592f7e104c2428f51947ddf6f80d6dac07938f0ac14", Started:(*bool)(nil)}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc00618b0a0), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", ImageID:"", ContainerID:"", Started:(*bool)(nil)}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc00618b080), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"registry.k8s.io/pause:3.8", ImageID:"", ContainerID:"", Started:(*bool)(0xc0040fba1f)}}, QOSClass:"Burstable", EphemeralContainerStatuses:[]v1.ContainerStatus(nil)}} -[AfterEach] [sig-node] InitContainer [NodeConformance] +STEP: Creating a kubernetes client 02/25/23 10:09:25.6 +Feb 25 10:09:25.600: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 10:09:25.602 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:09:25.655 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:09:25.659 +[It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:108 +STEP: Creating configMap with name projected-configmap-test-volume-map-76cf3840-1af4-4368-aa3e-45552dd89905 02/25/23 10:09:25.664 +STEP: Creating a pod to test consume configMaps 02/25/23 10:09:25.676 +Feb 25 10:09:25.690: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb" in namespace "projected-95" to be "Succeeded or Failed" +Feb 25 10:09:25.697: INFO: Pod "pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb": Phase="Pending", Reason="", readiness=false. Elapsed: 6.242802ms +Feb 25 10:09:27.707: INFO: Pod "pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016471413s +Feb 25 10:09:29.711: INFO: Pod "pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020617338s +STEP: Saw pod success 02/25/23 10:09:29.712 +Feb 25 10:09:29.712: INFO: Pod "pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb" satisfied condition "Succeeded or Failed" +Feb 25 10:09:29.721: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb container agnhost-container: +STEP: delete the pod 02/25/23 10:09:29.74 +Feb 25 10:09:29.767: INFO: Waiting for pod pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb to disappear +Feb 25 10:09:29.776: INFO: Pod pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb no longer exists +[AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 -Dec 14 16:01:07.091: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "init-container-3725" for this suite. 12/14/22 16:01:07.097 -{"msg":"PASSED [sig-node] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance]","completed":73,"skipped":1337,"failed":0} +Feb 25 10:09:29.777: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-95" for this suite. 02/25/23 10:09:29.789 +{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]","completed":75,"skipped":1532,"failed":0} ------------------------------ -• [SLOW TEST] [48.468 seconds] -[sig-node] InitContainer [NodeConformance] -test/e2e/common/node/framework.go:23 - should not start app containers if init containers fail on a RestartAlways pod [Conformance] - test/e2e/common/node/init_container.go:333 +• [4.206 seconds] +[sig-storage] Projected configMap +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:108 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] InitContainer [NodeConformance] + [BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:00:18.64 - Dec 14 16:00:18.640: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename init-container 12/14/22 16:00:18.648 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:00:18.731 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:00:18.735 - [BeforeEach] [sig-node] InitContainer [NodeConformance] - test/e2e/common/node/init_container.go:164 - [It] should not start app containers if init containers fail on a RestartAlways pod [Conformance] - test/e2e/common/node/init_container.go:333 - STEP: creating the pod 12/14/22 16:00:18.738 - Dec 14 16:00:18.739: INFO: PodSpec: initContainers in spec.initContainers - Dec 14 16:01:07.089: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-d2447cfc-008e-493a-81ec-e393d6b36751", GenerateName:"", Namespace:"init-container-3725", SelfLink:"", UID:"861d9d60-b513-4d48-b4c8-379788bd9ccd", ResourceVersion:"11774", Generation:0, CreationTimestamp:time.Date(2022, time.December, 14, 16, 0, 18, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"739103990"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:"e2e.test", Operation:"Update", APIVersion:"v1", Time:time.Date(2022, time.December, 14, 16, 0, 18, 0, time.Local), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc002e8cd80), Subresource:""}, v1.ManagedFieldsEntry{Manager:"kubelet", Operation:"Update", APIVersion:"v1", Time:time.Date(2022, time.December, 14, 16, 1, 7, 0, time.Local), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc002e8cdb0), Subresource:"status"}}}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"kube-api-access-gxpl2", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(nil), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(0xc00618b020), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil), Ephemeral:(*v1.EphemeralVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-api-access-gxpl2", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-api-access-gxpl2", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"registry.k8s.io/pause:3.8", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-api-access-gxpl2", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, EphemeralContainers:[]v1.EphemeralContainer(nil), RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc0040fb8d8), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"kiedai4iakae-2", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc006192e00), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0040fb970)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc0040fb990)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc0040fb998), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc0040fb99c), PreemptionPolicy:(*v1.PreemptionPolicy)(0xc0039dc920), Overhead:v1.ResourceList(nil), TopologySpreadConstraints:[]v1.TopologySpreadConstraint(nil), SetHostnameAsFQDN:(*bool)(nil), OS:(*v1.PodOS)(nil), HostUsers:(*bool)(nil)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2022, time.December, 14, 16, 0, 18, 0, time.Local), Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2022, time.December, 14, 16, 0, 18, 0, time.Local), Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2022, time.December, 14, 16, 0, 18, 0, time.Local), Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2022, time.December, 14, 16, 0, 18, 0, time.Local), Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"192.168.121.197", PodIP:"10.233.65.85", PodIPs:[]v1.PodIP{v1.PodIP{IP:"10.233.65.85"}}, StartTime:time.Date(2022, time.December, 14, 16, 0, 18, 0, time.Local), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc006192ee0)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc006192f50)}, Ready:false, RestartCount:3, Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", ImageID:"registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf", ContainerID:"cri-o://6b0d606803bf8a14fb2ce592f7e104c2428f51947ddf6f80d6dac07938f0ac14", Started:(*bool)(nil)}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc00618b0a0), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", ImageID:"", ContainerID:"", Started:(*bool)(nil)}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc00618b080), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"registry.k8s.io/pause:3.8", ImageID:"", ContainerID:"", Started:(*bool)(0xc0040fba1f)}}, QOSClass:"Burstable", EphemeralContainerStatuses:[]v1.ContainerStatus(nil)}} - [AfterEach] [sig-node] InitContainer [NodeConformance] + STEP: Creating a kubernetes client 02/25/23 10:09:25.6 + Feb 25 10:09:25.600: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 10:09:25.602 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:09:25.655 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:09:25.659 + [It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:108 + STEP: Creating configMap with name projected-configmap-test-volume-map-76cf3840-1af4-4368-aa3e-45552dd89905 02/25/23 10:09:25.664 + STEP: Creating a pod to test consume configMaps 02/25/23 10:09:25.676 + Feb 25 10:09:25.690: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb" in namespace "projected-95" to be "Succeeded or Failed" + Feb 25 10:09:25.697: INFO: Pod "pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb": Phase="Pending", Reason="", readiness=false. Elapsed: 6.242802ms + Feb 25 10:09:27.707: INFO: Pod "pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016471413s + Feb 25 10:09:29.711: INFO: Pod "pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.020617338s + STEP: Saw pod success 02/25/23 10:09:29.712 + Feb 25 10:09:29.712: INFO: Pod "pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb" satisfied condition "Succeeded or Failed" + Feb 25 10:09:29.721: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb container agnhost-container: + STEP: delete the pod 02/25/23 10:09:29.74 + Feb 25 10:09:29.767: INFO: Waiting for pod pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb to disappear + Feb 25 10:09:29.776: INFO: Pod pod-projected-configmaps-0cd3a136-8795-41a2-aaaa-336900873ecb no longer exists + [AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 - Dec 14 16:01:07.091: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "init-container-3725" for this suite. 12/14/22 16:01:07.097 + Feb 25 10:09:29.777: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-95" for this suite. 02/25/23 10:09:29.789 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SS ------------------------------ -[sig-network] IngressClass API - should support creating IngressClass API operations [Conformance] - test/e2e/network/ingressclass.go:223 -[BeforeEach] [sig-network] IngressClass API +[sig-apps] ReplicaSet + should validate Replicaset Status endpoints [Conformance] + test/e2e/apps/replica_set.go:176 +[BeforeEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:01:07.114 -Dec 14 16:01:07.114: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename ingressclass 12/14/22 16:01:07.12 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:01:07.147 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:01:07.151 -[BeforeEach] [sig-network] IngressClass API - test/e2e/network/ingressclass.go:211 -[It] should support creating IngressClass API operations [Conformance] - test/e2e/network/ingressclass.go:223 -STEP: getting /apis 12/14/22 16:01:07.155 -STEP: getting /apis/networking.k8s.io 12/14/22 16:01:07.159 -STEP: getting /apis/networking.k8s.iov1 12/14/22 16:01:07.161 -STEP: creating 12/14/22 16:01:07.162 -STEP: getting 12/14/22 16:01:07.194 -STEP: listing 12/14/22 16:01:07.199 -STEP: watching 12/14/22 16:01:07.204 -Dec 14 16:01:07.204: INFO: starting watch -STEP: patching 12/14/22 16:01:07.206 -STEP: updating 12/14/22 16:01:07.217 -Dec 14 16:01:07.234: INFO: waiting for watch events with expected annotations -Dec 14 16:01:07.234: INFO: saw patched and updated annotations -STEP: deleting 12/14/22 16:01:07.234 -STEP: deleting a collection 12/14/22 16:01:07.255 -[AfterEach] [sig-network] IngressClass API +STEP: Creating a kubernetes client 02/25/23 10:09:29.817 +Feb 25 10:09:29.818: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename replicaset 02/25/23 10:09:29.829 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:09:29.865 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:09:29.872 +[It] should validate Replicaset Status endpoints [Conformance] + test/e2e/apps/replica_set.go:176 +STEP: Create a Replicaset 02/25/23 10:09:29.887 +STEP: Verify that the required pods have come up. 02/25/23 10:09:29.902 +Feb 25 10:09:29.912: INFO: Pod name sample-pod: Found 0 pods out of 1 +Feb 25 10:09:34.927: INFO: Pod name sample-pod: Found 1 pods out of 1 +STEP: ensuring each pod is running 02/25/23 10:09:34.927 +STEP: Getting /status 02/25/23 10:09:34.928 +Feb 25 10:09:34.949: INFO: Replicaset test-rs has Conditions: [] +STEP: updating the Replicaset Status 02/25/23 10:09:34.949 +Feb 25 10:09:34.976: INFO: updatedStatus.Conditions: []v1.ReplicaSetCondition{v1.ReplicaSetCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} +STEP: watching for the ReplicaSet status to be updated 02/25/23 10:09:34.976 +Feb 25 10:09:34.983: INFO: Observed &ReplicaSet event: ADDED +Feb 25 10:09:34.984: INFO: Observed &ReplicaSet event: MODIFIED +Feb 25 10:09:34.985: INFO: Observed &ReplicaSet event: MODIFIED +Feb 25 10:09:34.986: INFO: Observed &ReplicaSet event: MODIFIED +Feb 25 10:09:34.986: INFO: Found replicaset test-rs in namespace replicaset-8923 with labels: map[name:sample-pod pod:httpd] annotations: map[] & Conditions: [{StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] +Feb 25 10:09:34.986: INFO: Replicaset test-rs has an updated status +STEP: patching the Replicaset Status 02/25/23 10:09:34.986 +Feb 25 10:09:34.987: INFO: Patch payload: {"status":{"conditions":[{"type":"StatusPatched","status":"True"}]}} +Feb 25 10:09:35.007: INFO: Patched status conditions: []v1.ReplicaSetCondition{v1.ReplicaSetCondition{Type:"StatusPatched", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"", Message:""}} +STEP: watching for the Replicaset status to be patched 02/25/23 10:09:35.007 +Feb 25 10:09:35.011: INFO: Observed &ReplicaSet event: ADDED +Feb 25 10:09:35.012: INFO: Observed &ReplicaSet event: MODIFIED +Feb 25 10:09:35.013: INFO: Observed &ReplicaSet event: MODIFIED +Feb 25 10:09:35.014: INFO: Observed &ReplicaSet event: MODIFIED +Feb 25 10:09:35.014: INFO: Observed replicaset test-rs in namespace replicaset-8923 with annotations: map[] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} +Feb 25 10:09:35.015: INFO: Observed &ReplicaSet event: MODIFIED +Feb 25 10:09:35.015: INFO: Found replicaset test-rs in namespace replicaset-8923 with labels: map[name:sample-pod pod:httpd] annotations: map[] & Conditions: {StatusPatched True 0001-01-01 00:00:00 +0000 UTC } +Feb 25 10:09:35.015: INFO: Replicaset test-rs has a patched status +[AfterEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:187 -Dec 14 16:01:07.288: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "ingressclass-7040" for this suite. 12/14/22 16:01:07.299 -{"msg":"PASSED [sig-network] IngressClass API should support creating IngressClass API operations [Conformance]","completed":74,"skipped":1372,"failed":0} +Feb 25 10:09:35.016: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "replicaset-8923" for this suite. 02/25/23 10:09:35.033 +{"msg":"PASSED [sig-apps] ReplicaSet should validate Replicaset Status endpoints [Conformance]","completed":76,"skipped":1534,"failed":0} ------------------------------ -• [0.205 seconds] -[sig-network] IngressClass API -test/e2e/network/common/framework.go:23 - should support creating IngressClass API operations [Conformance] - test/e2e/network/ingressclass.go:223 +• [SLOW TEST] [5.246 seconds] +[sig-apps] ReplicaSet +test/e2e/apps/framework.go:23 + should validate Replicaset Status endpoints [Conformance] + test/e2e/apps/replica_set.go:176 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] IngressClass API + [BeforeEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:01:07.114 - Dec 14 16:01:07.114: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename ingressclass 12/14/22 16:01:07.12 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:01:07.147 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:01:07.151 - [BeforeEach] [sig-network] IngressClass API - test/e2e/network/ingressclass.go:211 - [It] should support creating IngressClass API operations [Conformance] - test/e2e/network/ingressclass.go:223 - STEP: getting /apis 12/14/22 16:01:07.155 - STEP: getting /apis/networking.k8s.io 12/14/22 16:01:07.159 - STEP: getting /apis/networking.k8s.iov1 12/14/22 16:01:07.161 - STEP: creating 12/14/22 16:01:07.162 - STEP: getting 12/14/22 16:01:07.194 - STEP: listing 12/14/22 16:01:07.199 - STEP: watching 12/14/22 16:01:07.204 - Dec 14 16:01:07.204: INFO: starting watch - STEP: patching 12/14/22 16:01:07.206 - STEP: updating 12/14/22 16:01:07.217 - Dec 14 16:01:07.234: INFO: waiting for watch events with expected annotations - Dec 14 16:01:07.234: INFO: saw patched and updated annotations - STEP: deleting 12/14/22 16:01:07.234 - STEP: deleting a collection 12/14/22 16:01:07.255 - [AfterEach] [sig-network] IngressClass API + STEP: Creating a kubernetes client 02/25/23 10:09:29.817 + Feb 25 10:09:29.818: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename replicaset 02/25/23 10:09:29.829 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:09:29.865 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:09:29.872 + [It] should validate Replicaset Status endpoints [Conformance] + test/e2e/apps/replica_set.go:176 + STEP: Create a Replicaset 02/25/23 10:09:29.887 + STEP: Verify that the required pods have come up. 02/25/23 10:09:29.902 + Feb 25 10:09:29.912: INFO: Pod name sample-pod: Found 0 pods out of 1 + Feb 25 10:09:34.927: INFO: Pod name sample-pod: Found 1 pods out of 1 + STEP: ensuring each pod is running 02/25/23 10:09:34.927 + STEP: Getting /status 02/25/23 10:09:34.928 + Feb 25 10:09:34.949: INFO: Replicaset test-rs has Conditions: [] + STEP: updating the Replicaset Status 02/25/23 10:09:34.949 + Feb 25 10:09:34.976: INFO: updatedStatus.Conditions: []v1.ReplicaSetCondition{v1.ReplicaSetCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} + STEP: watching for the ReplicaSet status to be updated 02/25/23 10:09:34.976 + Feb 25 10:09:34.983: INFO: Observed &ReplicaSet event: ADDED + Feb 25 10:09:34.984: INFO: Observed &ReplicaSet event: MODIFIED + Feb 25 10:09:34.985: INFO: Observed &ReplicaSet event: MODIFIED + Feb 25 10:09:34.986: INFO: Observed &ReplicaSet event: MODIFIED + Feb 25 10:09:34.986: INFO: Found replicaset test-rs in namespace replicaset-8923 with labels: map[name:sample-pod pod:httpd] annotations: map[] & Conditions: [{StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] + Feb 25 10:09:34.986: INFO: Replicaset test-rs has an updated status + STEP: patching the Replicaset Status 02/25/23 10:09:34.986 + Feb 25 10:09:34.987: INFO: Patch payload: {"status":{"conditions":[{"type":"StatusPatched","status":"True"}]}} + Feb 25 10:09:35.007: INFO: Patched status conditions: []v1.ReplicaSetCondition{v1.ReplicaSetCondition{Type:"StatusPatched", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"", Message:""}} + STEP: watching for the Replicaset status to be patched 02/25/23 10:09:35.007 + Feb 25 10:09:35.011: INFO: Observed &ReplicaSet event: ADDED + Feb 25 10:09:35.012: INFO: Observed &ReplicaSet event: MODIFIED + Feb 25 10:09:35.013: INFO: Observed &ReplicaSet event: MODIFIED + Feb 25 10:09:35.014: INFO: Observed &ReplicaSet event: MODIFIED + Feb 25 10:09:35.014: INFO: Observed replicaset test-rs in namespace replicaset-8923 with annotations: map[] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} + Feb 25 10:09:35.015: INFO: Observed &ReplicaSet event: MODIFIED + Feb 25 10:09:35.015: INFO: Found replicaset test-rs in namespace replicaset-8923 with labels: map[name:sample-pod pod:httpd] annotations: map[] & Conditions: {StatusPatched True 0001-01-01 00:00:00 +0000 UTC } + Feb 25 10:09:35.015: INFO: Replicaset test-rs has a patched status + [AfterEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:187 - Dec 14 16:01:07.288: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "ingressclass-7040" for this suite. 12/14/22 16:01:07.299 + Feb 25 10:09:35.016: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "replicaset-8923" for this suite. 02/25/23 10:09:35.033 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSS ------------------------------ -[sig-node] Pods - should contain environment variables for services [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:443 -[BeforeEach] [sig-node] Pods +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + works for CRD with validation schema [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:68 +[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:01:07.328 -Dec 14 16:01:07.328: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename pods 12/14/22 16:01:07.331 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:01:07.358 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:01:07.362 -[BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 -[It] should contain environment variables for services [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:443 -Dec 14 16:01:07.378: INFO: Waiting up to 5m0s for pod "server-envvars-722af760-b2f2-4838-994b-8f45d7cdf07a" in namespace "pods-4015" to be "running and ready" -Dec 14 16:01:07.396: INFO: Pod "server-envvars-722af760-b2f2-4838-994b-8f45d7cdf07a": Phase="Pending", Reason="", readiness=false. Elapsed: 17.542828ms -Dec 14 16:01:07.396: INFO: The phase of Pod server-envvars-722af760-b2f2-4838-994b-8f45d7cdf07a is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:01:09.409: INFO: Pod "server-envvars-722af760-b2f2-4838-994b-8f45d7cdf07a": Phase="Running", Reason="", readiness=true. Elapsed: 2.03073044s -Dec 14 16:01:09.409: INFO: The phase of Pod server-envvars-722af760-b2f2-4838-994b-8f45d7cdf07a is Running (Ready = true) -Dec 14 16:01:09.409: INFO: Pod "server-envvars-722af760-b2f2-4838-994b-8f45d7cdf07a" satisfied condition "running and ready" -Dec 14 16:01:09.453: INFO: Waiting up to 5m0s for pod "client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe" in namespace "pods-4015" to be "Succeeded or Failed" -Dec 14 16:01:09.458: INFO: Pod "client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe": Phase="Pending", Reason="", readiness=false. Elapsed: 4.517439ms -Dec 14 16:01:11.464: INFO: Pod "client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010245784s -Dec 14 16:01:13.473: INFO: Pod "client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe": Phase="Pending", Reason="", readiness=false. Elapsed: 4.020104644s -Dec 14 16:01:15.470: INFO: Pod "client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016690337s -STEP: Saw pod success 12/14/22 16:01:15.47 -Dec 14 16:01:15.471: INFO: Pod "client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe" satisfied condition "Succeeded or Failed" -Dec 14 16:01:15.477: INFO: Trying to get logs from node kiedai4iakae-3 pod client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe container env3cont: -STEP: delete the pod 12/14/22 16:01:15.495 -Dec 14 16:01:15.520: INFO: Waiting for pod client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe to disappear -Dec 14 16:01:15.529: INFO: Pod client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe no longer exists -[AfterEach] [sig-node] Pods +STEP: Creating a kubernetes client 02/25/23 10:09:35.068 +Feb 25 10:09:35.068: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 10:09:35.074 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:09:35.123 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:09:35.131 +[It] works for CRD with validation schema [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:68 +Feb 25 10:09:35.138: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: kubectl validation (kubectl create and apply) allows request with known and required properties 02/25/23 10:09:42.402 +Feb 25 10:09:42.403: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 create -f -' +Feb 25 10:09:44.640: INFO: stderr: "" +Feb 25 10:09:44.641: INFO: stdout: "e2e-test-crd-publish-openapi-481-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" +Feb 25 10:09:44.642: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 delete e2e-test-crd-publish-openapi-481-crds test-foo' +Feb 25 10:09:44.820: INFO: stderr: "" +Feb 25 10:09:44.820: INFO: stdout: "e2e-test-crd-publish-openapi-481-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" +Feb 25 10:09:44.820: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 apply -f -' +Feb 25 10:09:46.408: INFO: stderr: "" +Feb 25 10:09:46.408: INFO: stdout: "e2e-test-crd-publish-openapi-481-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" +Feb 25 10:09:46.408: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 delete e2e-test-crd-publish-openapi-481-crds test-foo' +Feb 25 10:09:46.579: INFO: stderr: "" +Feb 25 10:09:46.579: INFO: stdout: "e2e-test-crd-publish-openapi-481-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" +STEP: kubectl validation (kubectl create and apply) rejects request with value outside defined enum values 02/25/23 10:09:46.579 +Feb 25 10:09:46.580: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 create -f -' +Feb 25 10:09:47.076: INFO: rc: 1 +STEP: kubectl validation (kubectl create and apply) rejects request with unknown properties when disallowed by the schema 02/25/23 10:09:47.076 +Feb 25 10:09:47.077: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 create -f -' +Feb 25 10:09:47.530: INFO: rc: 1 +Feb 25 10:09:47.530: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 apply -f -' +Feb 25 10:09:48.251: INFO: rc: 1 +STEP: kubectl validation (kubectl create and apply) rejects request without required properties 02/25/23 10:09:48.251 +Feb 25 10:09:48.252: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 create -f -' +Feb 25 10:09:48.758: INFO: rc: 1 +Feb 25 10:09:48.759: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 apply -f -' +Feb 25 10:09:49.376: INFO: rc: 1 +STEP: kubectl explain works to explain CR properties 02/25/23 10:09:49.376 +Feb 25 10:09:49.377: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 explain e2e-test-crd-publish-openapi-481-crds' +Feb 25 10:09:49.958: INFO: stderr: "" +Feb 25 10:09:49.958: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-481-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nDESCRIPTION:\n Foo CRD for Testing\n\nFIELDS:\n apiVersion\t\n APIVersion defines the versioned schema of this representation of an\n object. Servers should convert recognized schemas to the latest internal\n value, and may reject unrecognized values. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n kind\t\n Kind is a string value representing the REST resource this object\n represents. Servers may infer this from the endpoint the client submits\n requests to. Cannot be updated. In CamelCase. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n metadata\t\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n spec\t\n Specification of Foo\n\n status\t\n Status of Foo\n\n" +STEP: kubectl explain works to explain CR properties recursively 02/25/23 10:09:49.959 +Feb 25 10:09:49.961: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 explain e2e-test-crd-publish-openapi-481-crds.metadata' +Feb 25 10:09:50.569: INFO: stderr: "" +Feb 25 10:09:50.569: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-481-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: metadata \n\nDESCRIPTION:\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n ObjectMeta is metadata that all persisted resources must have, which\n includes all objects users must create.\n\nFIELDS:\n annotations\t\n Annotations is an unstructured key value map stored with a resource that\n may be set by external tools to store and retrieve arbitrary metadata. They\n are not queryable and should be preserved when modifying objects. More\n info: http://kubernetes.io/docs/user-guide/annotations\n\n creationTimestamp\t\n CreationTimestamp is a timestamp representing the server time when this\n object was created. It is not guaranteed to be set in happens-before order\n across separate operations. Clients may not set this value. It is\n represented in RFC3339 form and is in UTC.\n\n Populated by the system. Read-only. Null for lists. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n deletionGracePeriodSeconds\t\n Number of seconds allowed for this object to gracefully terminate before it\n will be removed from the system. Only set when deletionTimestamp is also\n set. May only be shortened. Read-only.\n\n deletionTimestamp\t\n DeletionTimestamp is RFC 3339 date and time at which this resource will be\n deleted. This field is set by the server when a graceful deletion is\n requested by the user, and is not directly settable by a client. The\n resource is expected to be deleted (no longer visible from resource lists,\n and not reachable by name) after the time in this field, once the\n finalizers list is empty. As long as the finalizers list contains items,\n deletion is blocked. Once the deletionTimestamp is set, this value may not\n be unset or be set further into the future, although it may be shortened or\n the resource may be deleted prior to this time. For example, a user may\n request that a pod is deleted in 30 seconds. The Kubelet will react by\n sending a graceful termination signal to the containers in the pod. After\n that 30 seconds, the Kubelet will send a hard termination signal (SIGKILL)\n to the container and after cleanup, remove the pod from the API. In the\n presence of network partitions, this object may still exist after this\n timestamp, until an administrator or automated process can determine the\n resource is fully terminated. If not set, graceful deletion of the object\n has not been requested.\n\n Populated by the system when a graceful deletion is requested. Read-only.\n More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n finalizers\t<[]string>\n Must be empty before the object is deleted from the registry. Each entry is\n an identifier for the responsible component that will remove the entry from\n the list. If the deletionTimestamp of the object is non-nil, entries in\n this list can only be removed. Finalizers may be processed and removed in\n any order. Order is NOT enforced because it introduces significant risk of\n stuck finalizers. finalizers is a shared field, any actor with permission\n can reorder it. If the finalizer list is processed in order, then this can\n lead to a situation in which the component responsible for the first\n finalizer in the list is waiting for a signal (field value, external\n system, or other) produced by a component responsible for a finalizer later\n in the list, resulting in a deadlock. Without enforced ordering finalizers\n are free to order amongst themselves and are not vulnerable to ordering\n changes in the list.\n\n generateName\t\n GenerateName is an optional prefix, used by the server, to generate a\n unique name ONLY IF the Name field has not been provided. If this field is\n used, the name returned to the client will be different than the name\n passed. This value will also be combined with a unique suffix. The provided\n value has the same validation rules as the Name field, and may be truncated\n by the length of the suffix required to make the value unique on the\n server.\n\n If this field is specified and the generated name exists, the server will\n return a 409.\n\n Applied only if Name is not specified. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency\n\n generation\t\n A sequence number representing a specific generation of the desired state.\n Populated by the system. Read-only.\n\n labels\t\n Map of string keys and values that can be used to organize and categorize\n (scope and select) objects. May match selectors of replication controllers\n and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n managedFields\t<[]Object>\n ManagedFields maps workflow-id and version to the set of fields that are\n managed by that workflow. This is mostly for internal housekeeping, and\n users typically shouldn't need to set or understand this field. A workflow\n can be the user's name, a controller's name, or the name of a specific\n apply path like \"ci-cd\". The set of fields is always in the version that\n the workflow used when modifying the object.\n\n name\t\n Name must be unique within a namespace. Is required when creating\n resources, although some resources may allow a client to request the\n generation of an appropriate name automatically. Name is primarily intended\n for creation idempotence and configuration definition. Cannot be updated.\n More info: http://kubernetes.io/docs/user-guide/identifiers#names\n\n namespace\t\n Namespace defines the space within which each name must be unique. An empty\n namespace is equivalent to the \"default\" namespace, but \"default\" is the\n canonical representation. Not all objects are required to be scoped to a\n namespace - the value of this field for those objects will be empty.\n\n Must be a DNS_LABEL. Cannot be updated. More info:\n http://kubernetes.io/docs/user-guide/namespaces\n\n ownerReferences\t<[]Object>\n List of objects depended by this object. If ALL objects in the list have\n been deleted, this object will be garbage collected. If this object is\n managed by a controller, then an entry in this list will point to this\n controller, with the controller field set to true. There cannot be more\n than one managing controller.\n\n resourceVersion\t\n An opaque value that represents the internal version of this object that\n can be used by clients to determine when objects have changed. May be used\n for optimistic concurrency, change detection, and the watch operation on a\n resource or set of resources. Clients must treat these values as opaque and\n passed unmodified back to the server. They may only be valid for a\n particular resource or set of resources.\n\n Populated by the system. Read-only. Value must be treated as opaque by\n clients and . More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency\n\n selfLink\t\n Deprecated: selfLink is a legacy read-only field that is no longer\n populated by the system.\n\n uid\t\n UID is the unique in time and space value for this object. It is typically\n generated by the server on successful creation of a resource and is not\n allowed to change on PUT operations.\n\n Populated by the system. Read-only. More info:\n http://kubernetes.io/docs/user-guide/identifiers#uids\n\n" +Feb 25 10:09:50.571: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 explain e2e-test-crd-publish-openapi-481-crds.spec' +Feb 25 10:09:51.073: INFO: stderr: "" +Feb 25 10:09:51.073: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-481-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: spec \n\nDESCRIPTION:\n Specification of Foo\n\nFIELDS:\n bars\t<[]Object>\n List of Bars and their specs.\n\n" +Feb 25 10:09:51.073: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 explain e2e-test-crd-publish-openapi-481-crds.spec.bars' +Feb 25 10:09:51.524: INFO: stderr: "" +Feb 25 10:09:51.524: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-481-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: bars <[]Object>\n\nDESCRIPTION:\n List of Bars and their specs.\n\nFIELDS:\n age\t\n Age of Bar.\n\n bazs\t<[]string>\n List of Bazs.\n\n feeling\t\n Whether Bar is feeling great.\n\n name\t -required-\n Name of Bar.\n\n" +STEP: kubectl explain works to return error when explain is called on property that doesn't exist 02/25/23 10:09:51.524 +Feb 25 10:09:51.525: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 explain e2e-test-crd-publish-openapi-481-crds.spec.bars2' +Feb 25 10:09:52.143: INFO: rc: 1 +[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:01:15.529: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-4015" for this suite. 12/14/22 16:01:15.537 -{"msg":"PASSED [sig-node] Pods should contain environment variables for services [NodeConformance] [Conformance]","completed":75,"skipped":1419,"failed":0} +Feb 25 10:09:59.165: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "crd-publish-openapi-7444" for this suite. 02/25/23 10:09:59.189 +{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD with validation schema [Conformance]","completed":77,"skipped":1544,"failed":0} ------------------------------ -• [SLOW TEST] [8.218 seconds] -[sig-node] Pods -test/e2e/common/node/framework.go:23 - should contain environment variables for services [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:443 +• [SLOW TEST] [24.139 seconds] +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + works for CRD with validation schema [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:68 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Pods + [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:01:07.328 - Dec 14 16:01:07.328: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename pods 12/14/22 16:01:07.331 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:01:07.358 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:01:07.362 - [BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 - [It] should contain environment variables for services [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:443 - Dec 14 16:01:07.378: INFO: Waiting up to 5m0s for pod "server-envvars-722af760-b2f2-4838-994b-8f45d7cdf07a" in namespace "pods-4015" to be "running and ready" - Dec 14 16:01:07.396: INFO: Pod "server-envvars-722af760-b2f2-4838-994b-8f45d7cdf07a": Phase="Pending", Reason="", readiness=false. Elapsed: 17.542828ms - Dec 14 16:01:07.396: INFO: The phase of Pod server-envvars-722af760-b2f2-4838-994b-8f45d7cdf07a is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:01:09.409: INFO: Pod "server-envvars-722af760-b2f2-4838-994b-8f45d7cdf07a": Phase="Running", Reason="", readiness=true. Elapsed: 2.03073044s - Dec 14 16:01:09.409: INFO: The phase of Pod server-envvars-722af760-b2f2-4838-994b-8f45d7cdf07a is Running (Ready = true) - Dec 14 16:01:09.409: INFO: Pod "server-envvars-722af760-b2f2-4838-994b-8f45d7cdf07a" satisfied condition "running and ready" - Dec 14 16:01:09.453: INFO: Waiting up to 5m0s for pod "client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe" in namespace "pods-4015" to be "Succeeded or Failed" - Dec 14 16:01:09.458: INFO: Pod "client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe": Phase="Pending", Reason="", readiness=false. Elapsed: 4.517439ms - Dec 14 16:01:11.464: INFO: Pod "client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010245784s - Dec 14 16:01:13.473: INFO: Pod "client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe": Phase="Pending", Reason="", readiness=false. Elapsed: 4.020104644s - Dec 14 16:01:15.470: INFO: Pod "client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016690337s - STEP: Saw pod success 12/14/22 16:01:15.47 - Dec 14 16:01:15.471: INFO: Pod "client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe" satisfied condition "Succeeded or Failed" - Dec 14 16:01:15.477: INFO: Trying to get logs from node kiedai4iakae-3 pod client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe container env3cont: - STEP: delete the pod 12/14/22 16:01:15.495 - Dec 14 16:01:15.520: INFO: Waiting for pod client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe to disappear - Dec 14 16:01:15.529: INFO: Pod client-envvars-8ced89ac-d7e0-4cb0-ade2-f86b9ec4e9fe no longer exists - [AfterEach] [sig-node] Pods + STEP: Creating a kubernetes client 02/25/23 10:09:35.068 + Feb 25 10:09:35.068: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 10:09:35.074 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:09:35.123 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:09:35.131 + [It] works for CRD with validation schema [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:68 + Feb 25 10:09:35.138: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: kubectl validation (kubectl create and apply) allows request with known and required properties 02/25/23 10:09:42.402 + Feb 25 10:09:42.403: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 create -f -' + Feb 25 10:09:44.640: INFO: stderr: "" + Feb 25 10:09:44.641: INFO: stdout: "e2e-test-crd-publish-openapi-481-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" + Feb 25 10:09:44.642: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 delete e2e-test-crd-publish-openapi-481-crds test-foo' + Feb 25 10:09:44.820: INFO: stderr: "" + Feb 25 10:09:44.820: INFO: stdout: "e2e-test-crd-publish-openapi-481-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" + Feb 25 10:09:44.820: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 apply -f -' + Feb 25 10:09:46.408: INFO: stderr: "" + Feb 25 10:09:46.408: INFO: stdout: "e2e-test-crd-publish-openapi-481-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" + Feb 25 10:09:46.408: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 delete e2e-test-crd-publish-openapi-481-crds test-foo' + Feb 25 10:09:46.579: INFO: stderr: "" + Feb 25 10:09:46.579: INFO: stdout: "e2e-test-crd-publish-openapi-481-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" + STEP: kubectl validation (kubectl create and apply) rejects request with value outside defined enum values 02/25/23 10:09:46.579 + Feb 25 10:09:46.580: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 create -f -' + Feb 25 10:09:47.076: INFO: rc: 1 + STEP: kubectl validation (kubectl create and apply) rejects request with unknown properties when disallowed by the schema 02/25/23 10:09:47.076 + Feb 25 10:09:47.077: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 create -f -' + Feb 25 10:09:47.530: INFO: rc: 1 + Feb 25 10:09:47.530: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 apply -f -' + Feb 25 10:09:48.251: INFO: rc: 1 + STEP: kubectl validation (kubectl create and apply) rejects request without required properties 02/25/23 10:09:48.251 + Feb 25 10:09:48.252: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 create -f -' + Feb 25 10:09:48.758: INFO: rc: 1 + Feb 25 10:09:48.759: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 --namespace=crd-publish-openapi-7444 apply -f -' + Feb 25 10:09:49.376: INFO: rc: 1 + STEP: kubectl explain works to explain CR properties 02/25/23 10:09:49.376 + Feb 25 10:09:49.377: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 explain e2e-test-crd-publish-openapi-481-crds' + Feb 25 10:09:49.958: INFO: stderr: "" + Feb 25 10:09:49.958: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-481-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nDESCRIPTION:\n Foo CRD for Testing\n\nFIELDS:\n apiVersion\t\n APIVersion defines the versioned schema of this representation of an\n object. Servers should convert recognized schemas to the latest internal\n value, and may reject unrecognized values. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n kind\t\n Kind is a string value representing the REST resource this object\n represents. Servers may infer this from the endpoint the client submits\n requests to. Cannot be updated. In CamelCase. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n metadata\t\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n spec\t\n Specification of Foo\n\n status\t\n Status of Foo\n\n" + STEP: kubectl explain works to explain CR properties recursively 02/25/23 10:09:49.959 + Feb 25 10:09:49.961: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 explain e2e-test-crd-publish-openapi-481-crds.metadata' + Feb 25 10:09:50.569: INFO: stderr: "" + Feb 25 10:09:50.569: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-481-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: metadata \n\nDESCRIPTION:\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n ObjectMeta is metadata that all persisted resources must have, which\n includes all objects users must create.\n\nFIELDS:\n annotations\t\n Annotations is an unstructured key value map stored with a resource that\n may be set by external tools to store and retrieve arbitrary metadata. They\n are not queryable and should be preserved when modifying objects. More\n info: http://kubernetes.io/docs/user-guide/annotations\n\n creationTimestamp\t\n CreationTimestamp is a timestamp representing the server time when this\n object was created. It is not guaranteed to be set in happens-before order\n across separate operations. Clients may not set this value. It is\n represented in RFC3339 form and is in UTC.\n\n Populated by the system. Read-only. Null for lists. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n deletionGracePeriodSeconds\t\n Number of seconds allowed for this object to gracefully terminate before it\n will be removed from the system. Only set when deletionTimestamp is also\n set. May only be shortened. Read-only.\n\n deletionTimestamp\t\n DeletionTimestamp is RFC 3339 date and time at which this resource will be\n deleted. This field is set by the server when a graceful deletion is\n requested by the user, and is not directly settable by a client. The\n resource is expected to be deleted (no longer visible from resource lists,\n and not reachable by name) after the time in this field, once the\n finalizers list is empty. As long as the finalizers list contains items,\n deletion is blocked. Once the deletionTimestamp is set, this value may not\n be unset or be set further into the future, although it may be shortened or\n the resource may be deleted prior to this time. For example, a user may\n request that a pod is deleted in 30 seconds. The Kubelet will react by\n sending a graceful termination signal to the containers in the pod. After\n that 30 seconds, the Kubelet will send a hard termination signal (SIGKILL)\n to the container and after cleanup, remove the pod from the API. In the\n presence of network partitions, this object may still exist after this\n timestamp, until an administrator or automated process can determine the\n resource is fully terminated. If not set, graceful deletion of the object\n has not been requested.\n\n Populated by the system when a graceful deletion is requested. Read-only.\n More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n finalizers\t<[]string>\n Must be empty before the object is deleted from the registry. Each entry is\n an identifier for the responsible component that will remove the entry from\n the list. If the deletionTimestamp of the object is non-nil, entries in\n this list can only be removed. Finalizers may be processed and removed in\n any order. Order is NOT enforced because it introduces significant risk of\n stuck finalizers. finalizers is a shared field, any actor with permission\n can reorder it. If the finalizer list is processed in order, then this can\n lead to a situation in which the component responsible for the first\n finalizer in the list is waiting for a signal (field value, external\n system, or other) produced by a component responsible for a finalizer later\n in the list, resulting in a deadlock. Without enforced ordering finalizers\n are free to order amongst themselves and are not vulnerable to ordering\n changes in the list.\n\n generateName\t\n GenerateName is an optional prefix, used by the server, to generate a\n unique name ONLY IF the Name field has not been provided. If this field is\n used, the name returned to the client will be different than the name\n passed. This value will also be combined with a unique suffix. The provided\n value has the same validation rules as the Name field, and may be truncated\n by the length of the suffix required to make the value unique on the\n server.\n\n If this field is specified and the generated name exists, the server will\n return a 409.\n\n Applied only if Name is not specified. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency\n\n generation\t\n A sequence number representing a specific generation of the desired state.\n Populated by the system. Read-only.\n\n labels\t\n Map of string keys and values that can be used to organize and categorize\n (scope and select) objects. May match selectors of replication controllers\n and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n managedFields\t<[]Object>\n ManagedFields maps workflow-id and version to the set of fields that are\n managed by that workflow. This is mostly for internal housekeeping, and\n users typically shouldn't need to set or understand this field. A workflow\n can be the user's name, a controller's name, or the name of a specific\n apply path like \"ci-cd\". The set of fields is always in the version that\n the workflow used when modifying the object.\n\n name\t\n Name must be unique within a namespace. Is required when creating\n resources, although some resources may allow a client to request the\n generation of an appropriate name automatically. Name is primarily intended\n for creation idempotence and configuration definition. Cannot be updated.\n More info: http://kubernetes.io/docs/user-guide/identifiers#names\n\n namespace\t\n Namespace defines the space within which each name must be unique. An empty\n namespace is equivalent to the \"default\" namespace, but \"default\" is the\n canonical representation. Not all objects are required to be scoped to a\n namespace - the value of this field for those objects will be empty.\n\n Must be a DNS_LABEL. Cannot be updated. More info:\n http://kubernetes.io/docs/user-guide/namespaces\n\n ownerReferences\t<[]Object>\n List of objects depended by this object. If ALL objects in the list have\n been deleted, this object will be garbage collected. If this object is\n managed by a controller, then an entry in this list will point to this\n controller, with the controller field set to true. There cannot be more\n than one managing controller.\n\n resourceVersion\t\n An opaque value that represents the internal version of this object that\n can be used by clients to determine when objects have changed. May be used\n for optimistic concurrency, change detection, and the watch operation on a\n resource or set of resources. Clients must treat these values as opaque and\n passed unmodified back to the server. They may only be valid for a\n particular resource or set of resources.\n\n Populated by the system. Read-only. Value must be treated as opaque by\n clients and . More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency\n\n selfLink\t\n Deprecated: selfLink is a legacy read-only field that is no longer\n populated by the system.\n\n uid\t\n UID is the unique in time and space value for this object. It is typically\n generated by the server on successful creation of a resource and is not\n allowed to change on PUT operations.\n\n Populated by the system. Read-only. More info:\n http://kubernetes.io/docs/user-guide/identifiers#uids\n\n" + Feb 25 10:09:50.571: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 explain e2e-test-crd-publish-openapi-481-crds.spec' + Feb 25 10:09:51.073: INFO: stderr: "" + Feb 25 10:09:51.073: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-481-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: spec \n\nDESCRIPTION:\n Specification of Foo\n\nFIELDS:\n bars\t<[]Object>\n List of Bars and their specs.\n\n" + Feb 25 10:09:51.073: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 explain e2e-test-crd-publish-openapi-481-crds.spec.bars' + Feb 25 10:09:51.524: INFO: stderr: "" + Feb 25 10:09:51.524: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-481-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: bars <[]Object>\n\nDESCRIPTION:\n List of Bars and their specs.\n\nFIELDS:\n age\t\n Age of Bar.\n\n bazs\t<[]string>\n List of Bazs.\n\n feeling\t\n Whether Bar is feeling great.\n\n name\t -required-\n Name of Bar.\n\n" + STEP: kubectl explain works to return error when explain is called on property that doesn't exist 02/25/23 10:09:51.524 + Feb 25 10:09:51.525: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-7444 explain e2e-test-crd-publish-openapi-481-crds.spec.bars2' + Feb 25 10:09:52.143: INFO: rc: 1 + [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:01:15.529: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "pods-4015" for this suite. 12/14/22 16:01:15.537 + Feb 25 10:09:59.165: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "crd-publish-openapi-7444" for this suite. 02/25/23 10:09:59.189 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSS +SSS ------------------------------ -[sig-storage] Downward API volume - should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:83 -[BeforeEach] [sig-storage] Downward API volume +[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] + should be able to convert from CR v1 to CR v2 [Conformance] + test/e2e/apimachinery/crd_conversion_webhook.go:149 +[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:01:15.554 -Dec 14 16:01:15.555: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 16:01:15.556 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:01:15.579 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:01:15.584 -[BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 -[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:83 -STEP: Creating a pod to test downward API volume plugin 12/14/22 16:01:15.593 -Dec 14 16:01:15.610: INFO: Waiting up to 5m0s for pod "downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927" in namespace "downward-api-3389" to be "Succeeded or Failed" -Dec 14 16:01:15.618: INFO: Pod "downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927": Phase="Pending", Reason="", readiness=false. Elapsed: 7.515168ms -Dec 14 16:01:17.630: INFO: Pod "downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019334127s -Dec 14 16:01:19.628: INFO: Pod "downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01781789s -STEP: Saw pod success 12/14/22 16:01:19.628 -Dec 14 16:01:19.629: INFO: Pod "downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927" satisfied condition "Succeeded or Failed" -Dec 14 16:01:19.636: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927 container client-container: -STEP: delete the pod 12/14/22 16:01:19.649 -Dec 14 16:01:19.670: INFO: Waiting for pod downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927 to disappear -Dec 14 16:01:19.674: INFO: Pod downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927 no longer exists -[AfterEach] [sig-storage] Downward API volume +STEP: Creating a kubernetes client 02/25/23 10:09:59.209 +Feb 25 10:09:59.209: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename crd-webhook 02/25/23 10:09:59.212 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:09:59.246 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:09:59.251 +[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/crd_conversion_webhook.go:128 +STEP: Setting up server cert 02/25/23 10:09:59.255 +STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication 02/25/23 10:10:00.859 +STEP: Deploying the custom resource conversion webhook pod 02/25/23 10:10:00.894 +STEP: Wait for the deployment to be ready 02/25/23 10:10:00.924 +Feb 25 10:10:00.941: INFO: new replicaset for deployment "sample-crd-conversion-webhook-deployment" is yet to be created +STEP: Deploying the webhook service 02/25/23 10:10:02.968 +STEP: Verifying the service has paired with the endpoint 02/25/23 10:10:02.993 +Feb 25 10:10:03.993: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1 +[It] should be able to convert from CR v1 to CR v2 [Conformance] + test/e2e/apimachinery/crd_conversion_webhook.go:149 +Feb 25 10:10:04.003: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Creating a v1 custom resource 02/25/23 10:10:06.864 +STEP: v2 custom resource should be converted 02/25/23 10:10:06.895 +[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:01:19.675: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-3389" for this suite. 12/14/22 16:01:19.68 -{"msg":"PASSED [sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","completed":76,"skipped":1435,"failed":0} +Feb 25 10:10:07.441: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "crd-webhook-4887" for this suite. 02/25/23 10:10:07.452 +[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/crd_conversion_webhook.go:139 +{"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert from CR v1 to CR v2 [Conformance]","completed":78,"skipped":1547,"failed":0} ------------------------------ -• [4.136 seconds] -[sig-storage] Downward API volume -test/e2e/common/storage/framework.go:23 - should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:83 +• [SLOW TEST] [8.363 seconds] +[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + should be able to convert from CR v1 to CR v2 [Conformance] + test/e2e/apimachinery/crd_conversion_webhook.go:149 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Downward API volume + [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:01:15.554 - Dec 14 16:01:15.555: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 16:01:15.556 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:01:15.579 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:01:15.584 - [BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 - [It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:83 - STEP: Creating a pod to test downward API volume plugin 12/14/22 16:01:15.593 - Dec 14 16:01:15.610: INFO: Waiting up to 5m0s for pod "downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927" in namespace "downward-api-3389" to be "Succeeded or Failed" - Dec 14 16:01:15.618: INFO: Pod "downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927": Phase="Pending", Reason="", readiness=false. Elapsed: 7.515168ms - Dec 14 16:01:17.630: INFO: Pod "downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019334127s - Dec 14 16:01:19.628: INFO: Pod "downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01781789s - STEP: Saw pod success 12/14/22 16:01:19.628 - Dec 14 16:01:19.629: INFO: Pod "downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927" satisfied condition "Succeeded or Failed" - Dec 14 16:01:19.636: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927 container client-container: - STEP: delete the pod 12/14/22 16:01:19.649 - Dec 14 16:01:19.670: INFO: Waiting for pod downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927 to disappear - Dec 14 16:01:19.674: INFO: Pod downwardapi-volume-1e0822a7-c57d-4d49-87e2-0f42c863f927 no longer exists - [AfterEach] [sig-storage] Downward API volume + STEP: Creating a kubernetes client 02/25/23 10:09:59.209 + Feb 25 10:09:59.209: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename crd-webhook 02/25/23 10:09:59.212 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:09:59.246 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:09:59.251 + [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/crd_conversion_webhook.go:128 + STEP: Setting up server cert 02/25/23 10:09:59.255 + STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication 02/25/23 10:10:00.859 + STEP: Deploying the custom resource conversion webhook pod 02/25/23 10:10:00.894 + STEP: Wait for the deployment to be ready 02/25/23 10:10:00.924 + Feb 25 10:10:00.941: INFO: new replicaset for deployment "sample-crd-conversion-webhook-deployment" is yet to be created + STEP: Deploying the webhook service 02/25/23 10:10:02.968 + STEP: Verifying the service has paired with the endpoint 02/25/23 10:10:02.993 + Feb 25 10:10:03.993: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1 + [It] should be able to convert from CR v1 to CR v2 [Conformance] + test/e2e/apimachinery/crd_conversion_webhook.go:149 + Feb 25 10:10:04.003: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Creating a v1 custom resource 02/25/23 10:10:06.864 + STEP: v2 custom resource should be converted 02/25/23 10:10:06.895 + [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:01:19.675: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-3389" for this suite. 12/14/22 16:01:19.68 + Feb 25 10:10:07.441: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "crd-webhook-4887" for this suite. 02/25/23 10:10:07.452 + [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/crd_conversion_webhook.go:139 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSS +SSSSSSSSSSS ------------------------------ -[sig-node] Probing container - should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:148 -[BeforeEach] [sig-node] Probing container +[sig-cli] Kubectl client Kubectl replace + should update a single-container pod's image [Conformance] + test/e2e/kubectl/kubectl.go:1745 +[BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:01:19.698 -Dec 14 16:01:19.699: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-probe 12/14/22 16:01:19.7 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:01:19.727 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:01:19.732 -[BeforeEach] [sig-node] Probing container - test/e2e/common/node/container_probe.go:59 -[It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:148 -STEP: Creating pod busybox-7a6dbc3e-2d13-4e80-a589-69bfba1b80cf in namespace container-probe-8528 12/14/22 16:01:19.737 -Dec 14 16:01:19.759: INFO: Waiting up to 5m0s for pod "busybox-7a6dbc3e-2d13-4e80-a589-69bfba1b80cf" in namespace "container-probe-8528" to be "not pending" -Dec 14 16:01:19.766: INFO: Pod "busybox-7a6dbc3e-2d13-4e80-a589-69bfba1b80cf": Phase="Pending", Reason="", readiness=false. Elapsed: 6.167585ms -Dec 14 16:01:21.781: INFO: Pod "busybox-7a6dbc3e-2d13-4e80-a589-69bfba1b80cf": Phase="Running", Reason="", readiness=true. Elapsed: 2.021590765s -Dec 14 16:01:21.781: INFO: Pod "busybox-7a6dbc3e-2d13-4e80-a589-69bfba1b80cf" satisfied condition "not pending" -Dec 14 16:01:21.781: INFO: Started pod busybox-7a6dbc3e-2d13-4e80-a589-69bfba1b80cf in namespace container-probe-8528 -STEP: checking the pod's current state and verifying that restartCount is present 12/14/22 16:01:21.781 -Dec 14 16:01:21.787: INFO: Initial restart count of pod busybox-7a6dbc3e-2d13-4e80-a589-69bfba1b80cf is 0 -STEP: deleting the pod 12/14/22 16:05:23.024 -[AfterEach] [sig-node] Probing container +STEP: Creating a kubernetes client 02/25/23 10:10:07.574 +Feb 25 10:10:07.574: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 10:10:07.591 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:10:07.694 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:10:07.7 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 +[BeforeEach] Kubectl replace + test/e2e/kubectl/kubectl.go:1732 +[It] should update a single-container pod's image [Conformance] + test/e2e/kubectl/kubectl.go:1745 +STEP: running the image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 02/25/23 10:10:07.725 +Feb 25 10:10:07.726: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6187 run e2e-test-httpd-pod --image=registry.k8s.io/e2e-test-images/httpd:2.4.38-2 --pod-running-timeout=2m0s --labels=run=e2e-test-httpd-pod' +Feb 25 10:10:08.102: INFO: stderr: "" +Feb 25 10:10:08.102: INFO: stdout: "pod/e2e-test-httpd-pod created\n" +STEP: verifying the pod e2e-test-httpd-pod is running 02/25/23 10:10:08.102 +STEP: verifying the pod e2e-test-httpd-pod was created 02/25/23 10:10:13.154 +Feb 25 10:10:13.155: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6187 get pod e2e-test-httpd-pod -o json' +Feb 25 10:10:13.343: INFO: stderr: "" +Feb 25 10:10:13.344: INFO: stdout: "{\n \"apiVersion\": \"v1\",\n \"kind\": \"Pod\",\n \"metadata\": {\n \"creationTimestamp\": \"2023-02-25T10:10:08Z\",\n \"labels\": {\n \"run\": \"e2e-test-httpd-pod\"\n },\n \"name\": \"e2e-test-httpd-pod\",\n \"namespace\": \"kubectl-6187\",\n \"resourceVersion\": \"15345\",\n \"uid\": \"b631c4dc-f0bc-44ab-8fcb-274130da5202\"\n },\n \"spec\": {\n \"containers\": [\n {\n \"image\": \"registry.k8s.io/e2e-test-images/httpd:2.4.38-2\",\n \"imagePullPolicy\": \"IfNotPresent\",\n \"name\": \"e2e-test-httpd-pod\",\n \"resources\": {},\n \"terminationMessagePath\": \"/dev/termination-log\",\n \"terminationMessagePolicy\": \"File\",\n \"volumeMounts\": [\n {\n \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n \"name\": \"kube-api-access-jq8ng\",\n \"readOnly\": true\n }\n ]\n }\n ],\n \"dnsPolicy\": \"ClusterFirst\",\n \"enableServiceLinks\": true,\n \"nodeName\": \"diek3dua9gay-3\",\n \"preemptionPolicy\": \"PreemptLowerPriority\",\n \"priority\": 0,\n \"restartPolicy\": \"Always\",\n \"schedulerName\": \"default-scheduler\",\n \"securityContext\": {},\n \"serviceAccount\": \"default\",\n \"serviceAccountName\": \"default\",\n \"terminationGracePeriodSeconds\": 30,\n \"tolerations\": [\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/not-ready\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n },\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/unreachable\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n }\n ],\n \"volumes\": [\n {\n \"name\": \"kube-api-access-jq8ng\",\n \"projected\": {\n \"defaultMode\": 420,\n \"sources\": [\n {\n \"serviceAccountToken\": {\n \"expirationSeconds\": 3607,\n \"path\": \"token\"\n }\n },\n {\n \"configMap\": {\n \"items\": [\n {\n \"key\": \"ca.crt\",\n \"path\": \"ca.crt\"\n }\n ],\n \"name\": \"kube-root-ca.crt\"\n }\n },\n {\n \"downwardAPI\": {\n \"items\": [\n {\n \"fieldRef\": {\n \"apiVersion\": \"v1\",\n \"fieldPath\": \"metadata.namespace\"\n },\n \"path\": \"namespace\"\n }\n ]\n }\n }\n ]\n }\n }\n ]\n },\n \"status\": {\n \"conditions\": [\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2023-02-25T10:10:08Z\",\n \"status\": \"True\",\n \"type\": \"Initialized\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2023-02-25T10:10:09Z\",\n \"status\": \"True\",\n \"type\": \"Ready\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2023-02-25T10:10:09Z\",\n \"status\": \"True\",\n \"type\": \"ContainersReady\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2023-02-25T10:10:08Z\",\n \"status\": \"True\",\n \"type\": \"PodScheduled\"\n }\n ],\n \"containerStatuses\": [\n {\n \"containerID\": \"cri-o://aaa496443bfa857b5f086fca5b0b3c9638e36316c4b30b2e273cb7861e60bea1\",\n \"image\": \"registry.k8s.io/e2e-test-images/httpd:2.4.38-2\",\n \"imageID\": \"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3\",\n \"lastState\": {},\n \"name\": \"e2e-test-httpd-pod\",\n \"ready\": true,\n \"restartCount\": 0,\n \"started\": true,\n \"state\": {\n \"running\": {\n \"startedAt\": \"2023-02-25T10:10:09Z\"\n }\n }\n }\n ],\n \"hostIP\": \"192.168.121.44\",\n \"phase\": \"Running\",\n \"podIP\": \"10.233.66.132\",\n \"podIPs\": [\n {\n \"ip\": \"10.233.66.132\"\n }\n ],\n \"qosClass\": \"BestEffort\",\n \"startTime\": \"2023-02-25T10:10:08Z\"\n }\n}\n" +STEP: replace the image in the pod 02/25/23 10:10:13.344 +Feb 25 10:10:13.345: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6187 replace -f -' +Feb 25 10:10:15.759: INFO: stderr: "" +Feb 25 10:10:15.759: INFO: stdout: "pod/e2e-test-httpd-pod replaced\n" +STEP: verifying the pod e2e-test-httpd-pod has the right image registry.k8s.io/e2e-test-images/busybox:1.29-2 02/25/23 10:10:15.759 +[AfterEach] Kubectl replace + test/e2e/kubectl/kubectl.go:1736 +Feb 25 10:10:15.771: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6187 delete pods e2e-test-httpd-pod' +Feb 25 10:10:17.655: INFO: stderr: "" +Feb 25 10:10:17.656: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" +[AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 -Dec 14 16:05:23.048: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-probe-8528" for this suite. 12/14/22 16:05:23.066 -{"msg":"PASSED [sig-node] Probing container should *not* be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","completed":77,"skipped":1445,"failed":0} +Feb 25 10:10:17.656: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-6187" for this suite. 02/25/23 10:10:17.665 +{"msg":"PASSED [sig-cli] Kubectl client Kubectl replace should update a single-container pod's image [Conformance]","completed":79,"skipped":1558,"failed":0} ------------------------------ -• [SLOW TEST] [243.381 seconds] -[sig-node] Probing container -test/e2e/common/node/framework.go:23 - should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:148 +• [SLOW TEST] [10.110 seconds] +[sig-cli] Kubectl client +test/e2e/kubectl/framework.go:23 + Kubectl replace + test/e2e/kubectl/kubectl.go:1729 + should update a single-container pod's image [Conformance] + test/e2e/kubectl/kubectl.go:1745 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Probing container + [BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:01:19.698 - Dec 14 16:01:19.699: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-probe 12/14/22 16:01:19.7 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:01:19.727 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:01:19.732 - [BeforeEach] [sig-node] Probing container - test/e2e/common/node/container_probe.go:59 - [It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:148 - STEP: Creating pod busybox-7a6dbc3e-2d13-4e80-a589-69bfba1b80cf in namespace container-probe-8528 12/14/22 16:01:19.737 - Dec 14 16:01:19.759: INFO: Waiting up to 5m0s for pod "busybox-7a6dbc3e-2d13-4e80-a589-69bfba1b80cf" in namespace "container-probe-8528" to be "not pending" - Dec 14 16:01:19.766: INFO: Pod "busybox-7a6dbc3e-2d13-4e80-a589-69bfba1b80cf": Phase="Pending", Reason="", readiness=false. Elapsed: 6.167585ms - Dec 14 16:01:21.781: INFO: Pod "busybox-7a6dbc3e-2d13-4e80-a589-69bfba1b80cf": Phase="Running", Reason="", readiness=true. Elapsed: 2.021590765s - Dec 14 16:01:21.781: INFO: Pod "busybox-7a6dbc3e-2d13-4e80-a589-69bfba1b80cf" satisfied condition "not pending" - Dec 14 16:01:21.781: INFO: Started pod busybox-7a6dbc3e-2d13-4e80-a589-69bfba1b80cf in namespace container-probe-8528 - STEP: checking the pod's current state and verifying that restartCount is present 12/14/22 16:01:21.781 - Dec 14 16:01:21.787: INFO: Initial restart count of pod busybox-7a6dbc3e-2d13-4e80-a589-69bfba1b80cf is 0 - STEP: deleting the pod 12/14/22 16:05:23.024 - [AfterEach] [sig-node] Probing container + STEP: Creating a kubernetes client 02/25/23 10:10:07.574 + Feb 25 10:10:07.574: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 10:10:07.591 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:10:07.694 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:10:07.7 + [BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 + [BeforeEach] Kubectl replace + test/e2e/kubectl/kubectl.go:1732 + [It] should update a single-container pod's image [Conformance] + test/e2e/kubectl/kubectl.go:1745 + STEP: running the image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 02/25/23 10:10:07.725 + Feb 25 10:10:07.726: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6187 run e2e-test-httpd-pod --image=registry.k8s.io/e2e-test-images/httpd:2.4.38-2 --pod-running-timeout=2m0s --labels=run=e2e-test-httpd-pod' + Feb 25 10:10:08.102: INFO: stderr: "" + Feb 25 10:10:08.102: INFO: stdout: "pod/e2e-test-httpd-pod created\n" + STEP: verifying the pod e2e-test-httpd-pod is running 02/25/23 10:10:08.102 + STEP: verifying the pod e2e-test-httpd-pod was created 02/25/23 10:10:13.154 + Feb 25 10:10:13.155: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6187 get pod e2e-test-httpd-pod -o json' + Feb 25 10:10:13.343: INFO: stderr: "" + Feb 25 10:10:13.344: INFO: stdout: "{\n \"apiVersion\": \"v1\",\n \"kind\": \"Pod\",\n \"metadata\": {\n \"creationTimestamp\": \"2023-02-25T10:10:08Z\",\n \"labels\": {\n \"run\": \"e2e-test-httpd-pod\"\n },\n \"name\": \"e2e-test-httpd-pod\",\n \"namespace\": \"kubectl-6187\",\n \"resourceVersion\": \"15345\",\n \"uid\": \"b631c4dc-f0bc-44ab-8fcb-274130da5202\"\n },\n \"spec\": {\n \"containers\": [\n {\n \"image\": \"registry.k8s.io/e2e-test-images/httpd:2.4.38-2\",\n \"imagePullPolicy\": \"IfNotPresent\",\n \"name\": \"e2e-test-httpd-pod\",\n \"resources\": {},\n \"terminationMessagePath\": \"/dev/termination-log\",\n \"terminationMessagePolicy\": \"File\",\n \"volumeMounts\": [\n {\n \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n \"name\": \"kube-api-access-jq8ng\",\n \"readOnly\": true\n }\n ]\n }\n ],\n \"dnsPolicy\": \"ClusterFirst\",\n \"enableServiceLinks\": true,\n \"nodeName\": \"diek3dua9gay-3\",\n \"preemptionPolicy\": \"PreemptLowerPriority\",\n \"priority\": 0,\n \"restartPolicy\": \"Always\",\n \"schedulerName\": \"default-scheduler\",\n \"securityContext\": {},\n \"serviceAccount\": \"default\",\n \"serviceAccountName\": \"default\",\n \"terminationGracePeriodSeconds\": 30,\n \"tolerations\": [\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/not-ready\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n },\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/unreachable\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n }\n ],\n \"volumes\": [\n {\n \"name\": \"kube-api-access-jq8ng\",\n \"projected\": {\n \"defaultMode\": 420,\n \"sources\": [\n {\n \"serviceAccountToken\": {\n \"expirationSeconds\": 3607,\n \"path\": \"token\"\n }\n },\n {\n \"configMap\": {\n \"items\": [\n {\n \"key\": \"ca.crt\",\n \"path\": \"ca.crt\"\n }\n ],\n \"name\": \"kube-root-ca.crt\"\n }\n },\n {\n \"downwardAPI\": {\n \"items\": [\n {\n \"fieldRef\": {\n \"apiVersion\": \"v1\",\n \"fieldPath\": \"metadata.namespace\"\n },\n \"path\": \"namespace\"\n }\n ]\n }\n }\n ]\n }\n }\n ]\n },\n \"status\": {\n \"conditions\": [\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2023-02-25T10:10:08Z\",\n \"status\": \"True\",\n \"type\": \"Initialized\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2023-02-25T10:10:09Z\",\n \"status\": \"True\",\n \"type\": \"Ready\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2023-02-25T10:10:09Z\",\n \"status\": \"True\",\n \"type\": \"ContainersReady\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2023-02-25T10:10:08Z\",\n \"status\": \"True\",\n \"type\": \"PodScheduled\"\n }\n ],\n \"containerStatuses\": [\n {\n \"containerID\": \"cri-o://aaa496443bfa857b5f086fca5b0b3c9638e36316c4b30b2e273cb7861e60bea1\",\n \"image\": \"registry.k8s.io/e2e-test-images/httpd:2.4.38-2\",\n \"imageID\": \"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3\",\n \"lastState\": {},\n \"name\": \"e2e-test-httpd-pod\",\n \"ready\": true,\n \"restartCount\": 0,\n \"started\": true,\n \"state\": {\n \"running\": {\n \"startedAt\": \"2023-02-25T10:10:09Z\"\n }\n }\n }\n ],\n \"hostIP\": \"192.168.121.44\",\n \"phase\": \"Running\",\n \"podIP\": \"10.233.66.132\",\n \"podIPs\": [\n {\n \"ip\": \"10.233.66.132\"\n }\n ],\n \"qosClass\": \"BestEffort\",\n \"startTime\": \"2023-02-25T10:10:08Z\"\n }\n}\n" + STEP: replace the image in the pod 02/25/23 10:10:13.344 + Feb 25 10:10:13.345: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6187 replace -f -' + Feb 25 10:10:15.759: INFO: stderr: "" + Feb 25 10:10:15.759: INFO: stdout: "pod/e2e-test-httpd-pod replaced\n" + STEP: verifying the pod e2e-test-httpd-pod has the right image registry.k8s.io/e2e-test-images/busybox:1.29-2 02/25/23 10:10:15.759 + [AfterEach] Kubectl replace + test/e2e/kubectl/kubectl.go:1736 + Feb 25 10:10:15.771: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6187 delete pods e2e-test-httpd-pod' + Feb 25 10:10:17.655: INFO: stderr: "" + Feb 25 10:10:17.656: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" + [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 16:05:23.048: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-probe-8528" for this suite. 12/14/22 16:05:23.066 + Feb 25 10:10:17.656: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-6187" for this suite. 02/25/23 10:10:17.665 << End Captured GinkgoWriter Output ------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ [sig-node] Variable Expansion - should allow substituting values in a volume subpath [Conformance] - test/e2e/common/node/expansion.go:111 + should succeed in writing subpaths in container [Slow] [Conformance] + test/e2e/common/node/expansion.go:296 [BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:05:23.084 -Dec 14 16:05:23.084: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename var-expansion 12/14/22 16:05:23.092 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:05:23.125 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:05:23.129 -[It] should allow substituting values in a volume subpath [Conformance] - test/e2e/common/node/expansion.go:111 -STEP: Creating a pod to test substitution in volume subpath 12/14/22 16:05:23.133 -Dec 14 16:05:23.161: INFO: Waiting up to 5m0s for pod "var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d" in namespace "var-expansion-7451" to be "Succeeded or Failed" -Dec 14 16:05:23.168: INFO: Pod "var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d": Phase="Pending", Reason="", readiness=false. Elapsed: 6.518713ms -Dec 14 16:05:25.197: INFO: Pod "var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.035917311s -Dec 14 16:05:27.178: INFO: Pod "var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016166602s -STEP: Saw pod success 12/14/22 16:05:27.178 -Dec 14 16:05:27.180: INFO: Pod "var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d" satisfied condition "Succeeded or Failed" -Dec 14 16:05:27.191: INFO: Trying to get logs from node kiedai4iakae-3 pod var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d container dapi-container: -STEP: delete the pod 12/14/22 16:05:27.229 -Dec 14 16:05:27.244: INFO: Waiting for pod var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d to disappear -Dec 14 16:05:27.248: INFO: Pod var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d no longer exists +STEP: Creating a kubernetes client 02/25/23 10:10:17.691 +Feb 25 10:10:17.691: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename var-expansion 02/25/23 10:10:17.693 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:10:17.73 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:10:17.736 +[It] should succeed in writing subpaths in container [Slow] [Conformance] + test/e2e/common/node/expansion.go:296 +STEP: creating the pod 02/25/23 10:10:17.741 +STEP: waiting for pod running 02/25/23 10:10:17.761 +Feb 25 10:10:17.762: INFO: Waiting up to 2m0s for pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f" in namespace "var-expansion-6210" to be "running" +Feb 25 10:10:17.774: INFO: Pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f": Phase="Pending", Reason="", readiness=false. Elapsed: 12.413336ms +Feb 25 10:10:19.782: INFO: Pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f": Phase="Running", Reason="", readiness=true. Elapsed: 2.020050004s +Feb 25 10:10:19.782: INFO: Pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f" satisfied condition "running" +STEP: creating a file in subpath 02/25/23 10:10:19.782 +Feb 25 10:10:19.788: INFO: ExecWithOptions {Command:[/bin/sh -c touch /volume_mount/mypath/foo/test.log] Namespace:var-expansion-6210 PodName:var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f ContainerName:dapi-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:10:19.789: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:10:19.790: INFO: ExecWithOptions: Clientset creation +Feb 25 10:10:19.790: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/var-expansion-6210/pods/var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f/exec?command=%2Fbin%2Fsh&command=-c&command=touch+%2Fvolume_mount%2Fmypath%2Ffoo%2Ftest.log&container=dapi-container&container=dapi-container&stderr=true&stdout=true) +STEP: test for file in mounted path 02/25/23 10:10:19.956 +Feb 25 10:10:19.963: INFO: ExecWithOptions {Command:[/bin/sh -c test -f /subpath_mount/test.log] Namespace:var-expansion-6210 PodName:var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f ContainerName:dapi-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:10:19.964: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:10:19.965: INFO: ExecWithOptions: Clientset creation +Feb 25 10:10:19.965: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/var-expansion-6210/pods/var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f/exec?command=%2Fbin%2Fsh&command=-c&command=test+-f+%2Fsubpath_mount%2Ftest.log&container=dapi-container&container=dapi-container&stderr=true&stdout=true) +STEP: updating the annotation value 02/25/23 10:10:20.076 +Feb 25 10:10:20.606: INFO: Successfully updated pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f" +STEP: waiting for annotated pod running 02/25/23 10:10:20.607 +Feb 25 10:10:20.607: INFO: Waiting up to 2m0s for pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f" in namespace "var-expansion-6210" to be "running" +Feb 25 10:10:20.613: INFO: Pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f": Phase="Running", Reason="", readiness=true. Elapsed: 5.475723ms +Feb 25 10:10:20.613: INFO: Pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f" satisfied condition "running" +STEP: deleting the pod gracefully 02/25/23 10:10:20.613 +Feb 25 10:10:20.614: INFO: Deleting pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f" in namespace "var-expansion-6210" +Feb 25 10:10:20.632: INFO: Wait up to 5m0s for pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f" to be fully deleted [AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 -Dec 14 16:05:27.249: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "var-expansion-7451" for this suite. 12/14/22 16:05:27.256 -{"msg":"PASSED [sig-node] Variable Expansion should allow substituting values in a volume subpath [Conformance]","completed":78,"skipped":1445,"failed":0} +Feb 25 10:10:54.653: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "var-expansion-6210" for this suite. 02/25/23 10:10:54.663 +{"msg":"PASSED [sig-node] Variable Expansion should succeed in writing subpaths in container [Slow] [Conformance]","completed":80,"skipped":1583,"failed":0} ------------------------------ -• [4.183 seconds] +• [SLOW TEST] [36.987 seconds] [sig-node] Variable Expansion test/e2e/common/node/framework.go:23 - should allow substituting values in a volume subpath [Conformance] - test/e2e/common/node/expansion.go:111 + should succeed in writing subpaths in container [Slow] [Conformance] + test/e2e/common/node/expansion.go:296 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:05:23.084 - Dec 14 16:05:23.084: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename var-expansion 12/14/22 16:05:23.092 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:05:23.125 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:05:23.129 - [It] should allow substituting values in a volume subpath [Conformance] - test/e2e/common/node/expansion.go:111 - STEP: Creating a pod to test substitution in volume subpath 12/14/22 16:05:23.133 - Dec 14 16:05:23.161: INFO: Waiting up to 5m0s for pod "var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d" in namespace "var-expansion-7451" to be "Succeeded or Failed" - Dec 14 16:05:23.168: INFO: Pod "var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d": Phase="Pending", Reason="", readiness=false. Elapsed: 6.518713ms - Dec 14 16:05:25.197: INFO: Pod "var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.035917311s - Dec 14 16:05:27.178: INFO: Pod "var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016166602s - STEP: Saw pod success 12/14/22 16:05:27.178 - Dec 14 16:05:27.180: INFO: Pod "var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d" satisfied condition "Succeeded or Failed" - Dec 14 16:05:27.191: INFO: Trying to get logs from node kiedai4iakae-3 pod var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d container dapi-container: - STEP: delete the pod 12/14/22 16:05:27.229 - Dec 14 16:05:27.244: INFO: Waiting for pod var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d to disappear - Dec 14 16:05:27.248: INFO: Pod var-expansion-252420ed-b0ae-46f5-a9fa-65dc9993cb7d no longer exists + STEP: Creating a kubernetes client 02/25/23 10:10:17.691 + Feb 25 10:10:17.691: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename var-expansion 02/25/23 10:10:17.693 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:10:17.73 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:10:17.736 + [It] should succeed in writing subpaths in container [Slow] [Conformance] + test/e2e/common/node/expansion.go:296 + STEP: creating the pod 02/25/23 10:10:17.741 + STEP: waiting for pod running 02/25/23 10:10:17.761 + Feb 25 10:10:17.762: INFO: Waiting up to 2m0s for pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f" in namespace "var-expansion-6210" to be "running" + Feb 25 10:10:17.774: INFO: Pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f": Phase="Pending", Reason="", readiness=false. Elapsed: 12.413336ms + Feb 25 10:10:19.782: INFO: Pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f": Phase="Running", Reason="", readiness=true. Elapsed: 2.020050004s + Feb 25 10:10:19.782: INFO: Pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f" satisfied condition "running" + STEP: creating a file in subpath 02/25/23 10:10:19.782 + Feb 25 10:10:19.788: INFO: ExecWithOptions {Command:[/bin/sh -c touch /volume_mount/mypath/foo/test.log] Namespace:var-expansion-6210 PodName:var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f ContainerName:dapi-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:10:19.789: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:10:19.790: INFO: ExecWithOptions: Clientset creation + Feb 25 10:10:19.790: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/var-expansion-6210/pods/var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f/exec?command=%2Fbin%2Fsh&command=-c&command=touch+%2Fvolume_mount%2Fmypath%2Ffoo%2Ftest.log&container=dapi-container&container=dapi-container&stderr=true&stdout=true) + STEP: test for file in mounted path 02/25/23 10:10:19.956 + Feb 25 10:10:19.963: INFO: ExecWithOptions {Command:[/bin/sh -c test -f /subpath_mount/test.log] Namespace:var-expansion-6210 PodName:var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f ContainerName:dapi-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:10:19.964: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:10:19.965: INFO: ExecWithOptions: Clientset creation + Feb 25 10:10:19.965: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/var-expansion-6210/pods/var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f/exec?command=%2Fbin%2Fsh&command=-c&command=test+-f+%2Fsubpath_mount%2Ftest.log&container=dapi-container&container=dapi-container&stderr=true&stdout=true) + STEP: updating the annotation value 02/25/23 10:10:20.076 + Feb 25 10:10:20.606: INFO: Successfully updated pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f" + STEP: waiting for annotated pod running 02/25/23 10:10:20.607 + Feb 25 10:10:20.607: INFO: Waiting up to 2m0s for pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f" in namespace "var-expansion-6210" to be "running" + Feb 25 10:10:20.613: INFO: Pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f": Phase="Running", Reason="", readiness=true. Elapsed: 5.475723ms + Feb 25 10:10:20.613: INFO: Pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f" satisfied condition "running" + STEP: deleting the pod gracefully 02/25/23 10:10:20.613 + Feb 25 10:10:20.614: INFO: Deleting pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f" in namespace "var-expansion-6210" + Feb 25 10:10:20.632: INFO: Wait up to 5m0s for pod "var-expansion-f9145da5-c581-4035-8a6c-ff2fd561646f" to be fully deleted [AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 - Dec 14 16:05:27.249: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "var-expansion-7451" for this suite. 12/14/22 16:05:27.256 - << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-network] Services - should be able to change the type from ExternalName to NodePort [Conformance] - test/e2e/network/service.go:1443 -[BeforeEach] [sig-network] Services - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:05:27.274 -Dec 14 16:05:27.274: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 16:05:27.277 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:05:27.305 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:05:27.311 -[BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 -[It] should be able to change the type from ExternalName to NodePort [Conformance] - test/e2e/network/service.go:1443 -STEP: creating a service externalname-service with the type=ExternalName in namespace services-5268 12/14/22 16:05:27.315 -STEP: changing the ExternalName service to type=NodePort 12/14/22 16:05:27.327 -STEP: creating replication controller externalname-service in namespace services-5268 12/14/22 16:05:27.361 -I1214 16:05:27.381157 14 runners.go:193] Created replication controller with name: externalname-service, namespace: services-5268, replica count: 2 -I1214 16:05:30.432952 14 runners.go:193] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -Dec 14 16:05:30.433: INFO: Creating new exec pod -Dec 14 16:05:30.451: INFO: Waiting up to 5m0s for pod "execpodmhrp8" in namespace "services-5268" to be "running" -Dec 14 16:05:30.457: INFO: Pod "execpodmhrp8": Phase="Pending", Reason="", readiness=false. Elapsed: 6.428754ms -Dec 14 16:05:32.465: INFO: Pod "execpodmhrp8": Phase="Running", Reason="", readiness=true. Elapsed: 2.014448498s -Dec 14 16:05:32.465: INFO: Pod "execpodmhrp8" satisfied condition "running" -Dec 14 16:05:33.474: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 externalname-service 80' -Dec 14 16:05:33.751: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n" -Dec 14 16:05:33.751: INFO: stdout: "" -Dec 14 16:05:34.752: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 externalname-service 80' -Dec 14 16:05:35.021: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n" -Dec 14 16:05:35.021: INFO: stdout: "externalname-service-nrslf" -Dec 14 16:05:35.021: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.36.200 80' -Dec 14 16:05:35.234: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.36.200 80\nConnection to 10.233.36.200 80 port [tcp/http] succeeded!\n" -Dec 14 16:05:35.234: INFO: stdout: "" -Dec 14 16:05:36.236: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.36.200 80' -Dec 14 16:05:36.507: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.36.200 80\nConnection to 10.233.36.200 80 port [tcp/http] succeeded!\n" -Dec 14 16:05:36.507: INFO: stdout: "externalname-service-nrslf" -Dec 14 16:05:36.508: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.197 32280' -Dec 14 16:05:36.763: INFO: stderr: "+ + nc -v -t -w 2echo 192.168.121.197 32280\n hostName\nConnection to 192.168.121.197 32280 port [tcp/*] succeeded!\n" -Dec 14 16:05:36.763: INFO: stdout: "" -Dec 14 16:05:37.764: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.197 32280' -Dec 14 16:05:38.082: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.197 32280\nConnection to 192.168.121.197 32280 port [tcp/*] succeeded!\n" -Dec 14 16:05:38.082: INFO: stdout: "externalname-service-7vnr4" -Dec 14 16:05:38.083: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.249 32280' -Dec 14 16:05:38.423: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.249 32280\nConnection to 192.168.121.249 32280 port [tcp/*] succeeded!\n" -Dec 14 16:05:38.423: INFO: stdout: "" -Dec 14 16:05:39.424: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.249 32280' -Dec 14 16:05:39.614: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.249 32280\nConnection to 192.168.121.249 32280 port [tcp/*] succeeded!\n" -Dec 14 16:05:39.614: INFO: stdout: "" -Dec 14 16:05:40.424: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.249 32280' -Dec 14 16:05:40.633: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.249 32280\nConnection to 192.168.121.249 32280 port [tcp/*] succeeded!\n" -Dec 14 16:05:40.633: INFO: stdout: "externalname-service-7vnr4" -Dec 14 16:05:40.633: INFO: Cleaning up the ExternalName to NodePort test service -[AfterEach] [sig-network] Services - test/e2e/framework/framework.go:187 -Dec 14 16:05:40.674: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-5268" for this suite. 12/14/22 16:05:40.682 -[AfterEach] [sig-network] Services - test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to NodePort [Conformance]","completed":79,"skipped":1474,"failed":0} ------------------------------- -• [SLOW TEST] [13.421 seconds] -[sig-network] Services -test/e2e/network/common/framework.go:23 - should be able to change the type from ExternalName to NodePort [Conformance] - test/e2e/network/service.go:1443 - - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Services - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:05:27.274 - Dec 14 16:05:27.274: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 16:05:27.277 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:05:27.305 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:05:27.311 - [BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 - [It] should be able to change the type from ExternalName to NodePort [Conformance] - test/e2e/network/service.go:1443 - STEP: creating a service externalname-service with the type=ExternalName in namespace services-5268 12/14/22 16:05:27.315 - STEP: changing the ExternalName service to type=NodePort 12/14/22 16:05:27.327 - STEP: creating replication controller externalname-service in namespace services-5268 12/14/22 16:05:27.361 - I1214 16:05:27.381157 14 runners.go:193] Created replication controller with name: externalname-service, namespace: services-5268, replica count: 2 - I1214 16:05:30.432952 14 runners.go:193] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - Dec 14 16:05:30.433: INFO: Creating new exec pod - Dec 14 16:05:30.451: INFO: Waiting up to 5m0s for pod "execpodmhrp8" in namespace "services-5268" to be "running" - Dec 14 16:05:30.457: INFO: Pod "execpodmhrp8": Phase="Pending", Reason="", readiness=false. Elapsed: 6.428754ms - Dec 14 16:05:32.465: INFO: Pod "execpodmhrp8": Phase="Running", Reason="", readiness=true. Elapsed: 2.014448498s - Dec 14 16:05:32.465: INFO: Pod "execpodmhrp8" satisfied condition "running" - Dec 14 16:05:33.474: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 externalname-service 80' - Dec 14 16:05:33.751: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n" - Dec 14 16:05:33.751: INFO: stdout: "" - Dec 14 16:05:34.752: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 externalname-service 80' - Dec 14 16:05:35.021: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n" - Dec 14 16:05:35.021: INFO: stdout: "externalname-service-nrslf" - Dec 14 16:05:35.021: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.36.200 80' - Dec 14 16:05:35.234: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.36.200 80\nConnection to 10.233.36.200 80 port [tcp/http] succeeded!\n" - Dec 14 16:05:35.234: INFO: stdout: "" - Dec 14 16:05:36.236: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.36.200 80' - Dec 14 16:05:36.507: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.36.200 80\nConnection to 10.233.36.200 80 port [tcp/http] succeeded!\n" - Dec 14 16:05:36.507: INFO: stdout: "externalname-service-nrslf" - Dec 14 16:05:36.508: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.197 32280' - Dec 14 16:05:36.763: INFO: stderr: "+ + nc -v -t -w 2echo 192.168.121.197 32280\n hostName\nConnection to 192.168.121.197 32280 port [tcp/*] succeeded!\n" - Dec 14 16:05:36.763: INFO: stdout: "" - Dec 14 16:05:37.764: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.197 32280' - Dec 14 16:05:38.082: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.197 32280\nConnection to 192.168.121.197 32280 port [tcp/*] succeeded!\n" - Dec 14 16:05:38.082: INFO: stdout: "externalname-service-7vnr4" - Dec 14 16:05:38.083: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.249 32280' - Dec 14 16:05:38.423: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.249 32280\nConnection to 192.168.121.249 32280 port [tcp/*] succeeded!\n" - Dec 14 16:05:38.423: INFO: stdout: "" - Dec 14 16:05:39.424: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.249 32280' - Dec 14 16:05:39.614: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.249 32280\nConnection to 192.168.121.249 32280 port [tcp/*] succeeded!\n" - Dec 14 16:05:39.614: INFO: stdout: "" - Dec 14 16:05:40.424: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5268 exec execpodmhrp8 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.249 32280' - Dec 14 16:05:40.633: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.249 32280\nConnection to 192.168.121.249 32280 port [tcp/*] succeeded!\n" - Dec 14 16:05:40.633: INFO: stdout: "externalname-service-7vnr4" - Dec 14 16:05:40.633: INFO: Cleaning up the ExternalName to NodePort test service - [AfterEach] [sig-network] Services - test/e2e/framework/framework.go:187 - Dec 14 16:05:40.674: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-5268" for this suite. 12/14/22 16:05:40.682 - [AfterEach] [sig-network] Services - test/e2e/network/service.go:762 + Feb 25 10:10:54.653: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "var-expansion-6210" for this suite. 02/25/23 10:10:54.663 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - patching/updating a mutating webhook should work [Conformance] - test/e2e/apimachinery/webhook.go:507 + should mutate pod and apply defaults after mutation [Conformance] + test/e2e/apimachinery/webhook.go:263 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:05:40.696 -Dec 14 16:05:40.696: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 16:05:40.699 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:05:40.726 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:05:40.73 +STEP: Creating a kubernetes client 02/25/23 10:10:54.684 +Feb 25 10:10:54.684: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 10:10:54.688 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:10:54.727 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:10:54.733 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 16:05:40.754 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:05:41.948 -STEP: Deploying the webhook pod 12/14/22 16:05:41.964 -STEP: Wait for the deployment to be ready 12/14/22 16:05:41.987 -Dec 14 16:05:42.003: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set -STEP: Deploying the webhook service 12/14/22 16:05:44.042 -STEP: Verifying the service has paired with the endpoint 12/14/22 16:05:44.062 -Dec 14 16:05:45.062: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] patching/updating a mutating webhook should work [Conformance] - test/e2e/apimachinery/webhook.go:507 -STEP: Creating a mutating webhook configuration 12/14/22 16:05:45.07 -Dec 14 16:05:45.114: INFO: Waiting for webhook configuration to be ready... -STEP: Updating a mutating webhook configuration's rules to not include the create operation 12/14/22 16:05:45.239 -STEP: Creating a configMap that should not be mutated 12/14/22 16:05:45.25 -STEP: Patching a mutating webhook configuration's rules to include the create operation 12/14/22 16:05:45.269 -STEP: Creating a configMap that should be mutated 12/14/22 16:05:45.279 +STEP: Setting up server cert 02/25/23 10:10:54.764 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:10:55.246 +STEP: Deploying the webhook pod 02/25/23 10:10:55.264 +STEP: Wait for the deployment to be ready 02/25/23 10:10:55.291 +Feb 25 10:10:55.306: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set +STEP: Deploying the webhook service 02/25/23 10:10:57.333 +STEP: Verifying the service has paired with the endpoint 02/25/23 10:10:57.362 +Feb 25 10:10:58.362: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should mutate pod and apply defaults after mutation [Conformance] + test/e2e/apimachinery/webhook.go:263 +STEP: Registering the mutating pod webhook via the AdmissionRegistration API 02/25/23 10:10:58.373 +STEP: create a pod that should be updated by the webhook 02/25/23 10:10:58.418 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:05:45.441: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-8395" for this suite. 12/14/22 16:05:45.449 -STEP: Destroying namespace "webhook-8395-markers" for this suite. 12/14/22 16:05:45.457 +Feb 25 10:10:58.506: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-6493" for this suite. 02/25/23 10:10:58.52 +STEP: Destroying namespace "webhook-6493-markers" for this suite. 02/25/23 10:10:58.555 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a mutating webhook should work [Conformance]","completed":80,"skipped":1491,"failed":0} +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate pod and apply defaults after mutation [Conformance]","completed":81,"skipped":1605,"failed":0} ------------------------------ -• [4.825 seconds] +• [4.067 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/framework.go:23 - patching/updating a mutating webhook should work [Conformance] - test/e2e/apimachinery/webhook.go:507 + should mutate pod and apply defaults after mutation [Conformance] + test/e2e/apimachinery/webhook.go:263 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:05:40.696 - Dec 14 16:05:40.696: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 16:05:40.699 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:05:40.726 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:05:40.73 + STEP: Creating a kubernetes client 02/25/23 10:10:54.684 + Feb 25 10:10:54.684: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 10:10:54.688 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:10:54.727 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:10:54.733 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 16:05:40.754 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:05:41.948 - STEP: Deploying the webhook pod 12/14/22 16:05:41.964 - STEP: Wait for the deployment to be ready 12/14/22 16:05:41.987 - Dec 14 16:05:42.003: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set - STEP: Deploying the webhook service 12/14/22 16:05:44.042 - STEP: Verifying the service has paired with the endpoint 12/14/22 16:05:44.062 - Dec 14 16:05:45.062: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] patching/updating a mutating webhook should work [Conformance] - test/e2e/apimachinery/webhook.go:507 - STEP: Creating a mutating webhook configuration 12/14/22 16:05:45.07 - Dec 14 16:05:45.114: INFO: Waiting for webhook configuration to be ready... - STEP: Updating a mutating webhook configuration's rules to not include the create operation 12/14/22 16:05:45.239 - STEP: Creating a configMap that should not be mutated 12/14/22 16:05:45.25 - STEP: Patching a mutating webhook configuration's rules to include the create operation 12/14/22 16:05:45.269 - STEP: Creating a configMap that should be mutated 12/14/22 16:05:45.279 + STEP: Setting up server cert 02/25/23 10:10:54.764 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:10:55.246 + STEP: Deploying the webhook pod 02/25/23 10:10:55.264 + STEP: Wait for the deployment to be ready 02/25/23 10:10:55.291 + Feb 25 10:10:55.306: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set + STEP: Deploying the webhook service 02/25/23 10:10:57.333 + STEP: Verifying the service has paired with the endpoint 02/25/23 10:10:57.362 + Feb 25 10:10:58.362: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] should mutate pod and apply defaults after mutation [Conformance] + test/e2e/apimachinery/webhook.go:263 + STEP: Registering the mutating pod webhook via the AdmissionRegistration API 02/25/23 10:10:58.373 + STEP: create a pod that should be updated by the webhook 02/25/23 10:10:58.418 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:05:45.441: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-8395" for this suite. 12/14/22 16:05:45.449 - STEP: Destroying namespace "webhook-8395-markers" for this suite. 12/14/22 16:05:45.457 + Feb 25 10:10:58.506: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-6493" for this suite. 02/25/23 10:10:58.52 + STEP: Destroying namespace "webhook-6493-markers" for this suite. 02/25/23 10:10:58.555 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:104 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Secrets - optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:204 -[BeforeEach] [sig-storage] Secrets +[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] + should include custom resource definition resources in discovery documents [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:198 +[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:05:45.524 -Dec 14 16:05:45.525: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename secrets 12/14/22 16:05:45.534 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:05:45.567 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:05:45.571 -[It] optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:204 -STEP: Creating secret with name s-test-opt-del-381c6675-cab3-4aff-b10a-0d1be7b17aca 12/14/22 16:05:45.58 -STEP: Creating secret with name s-test-opt-upd-4bbaaf00-5a3b-49af-a40b-16b12aeed372 12/14/22 16:05:45.586 -STEP: Creating the pod 12/14/22 16:05:45.595 -Dec 14 16:05:45.609: INFO: Waiting up to 5m0s for pod "pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281" in namespace "secrets-1497" to be "running and ready" -Dec 14 16:05:45.615: INFO: Pod "pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281": Phase="Pending", Reason="", readiness=false. Elapsed: 6.586148ms -Dec 14 16:05:45.615: INFO: The phase of Pod pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:05:47.623: INFO: Pod "pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01452916s -Dec 14 16:05:47.623: INFO: The phase of Pod pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:05:49.629: INFO: Pod "pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281": Phase="Running", Reason="", readiness=true. Elapsed: 4.020093228s -Dec 14 16:05:49.629: INFO: The phase of Pod pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281 is Running (Ready = true) -Dec 14 16:05:49.629: INFO: Pod "pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281" satisfied condition "running and ready" -STEP: Deleting secret s-test-opt-del-381c6675-cab3-4aff-b10a-0d1be7b17aca 12/14/22 16:05:49.686 -STEP: Updating secret s-test-opt-upd-4bbaaf00-5a3b-49af-a40b-16b12aeed372 12/14/22 16:05:49.698 -STEP: Creating secret with name s-test-opt-create-505b30ff-fe09-4f13-b265-00d433538c8d 12/14/22 16:05:49.706 -STEP: waiting to observe update in volume 12/14/22 16:05:49.715 -[AfterEach] [sig-storage] Secrets +STEP: Creating a kubernetes client 02/25/23 10:10:58.797 +Feb 25 10:10:58.798: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename custom-resource-definition 02/25/23 10:10:58.805 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:10:58.852 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:10:58.858 +[It] should include custom resource definition resources in discovery documents [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:198 +STEP: fetching the /apis discovery document 02/25/23 10:10:58.865 +STEP: finding the apiextensions.k8s.io API group in the /apis discovery document 02/25/23 10:10:58.869 +STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis discovery document 02/25/23 10:10:58.87 +STEP: fetching the /apis/apiextensions.k8s.io discovery document 02/25/23 10:10:58.87 +STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis/apiextensions.k8s.io discovery document 02/25/23 10:10:58.873 +STEP: fetching the /apis/apiextensions.k8s.io/v1 discovery document 02/25/23 10:10:58.874 +STEP: finding customresourcedefinitions resources in the /apis/apiextensions.k8s.io/v1 discovery document 02/25/23 10:10:58.877 +[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:07:02.584: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-1497" for this suite. 12/14/22 16:07:02.592 -{"msg":"PASSED [sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance]","completed":81,"skipped":1498,"failed":0} +Feb 25 10:10:58.879: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "custom-resource-definition-8987" for this suite. 02/25/23 10:10:58.896 +{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] should include custom resource definition resources in discovery documents [Conformance]","completed":82,"skipped":1635,"failed":0} ------------------------------ -• [SLOW TEST] [77.078 seconds] -[sig-storage] Secrets -test/e2e/common/storage/framework.go:23 - optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:204 +• [0.147 seconds] +[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + should include custom resource definition resources in discovery documents [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:198 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Secrets + [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:05:45.524 - Dec 14 16:05:45.525: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename secrets 12/14/22 16:05:45.534 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:05:45.567 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:05:45.571 - [It] optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:204 - STEP: Creating secret with name s-test-opt-del-381c6675-cab3-4aff-b10a-0d1be7b17aca 12/14/22 16:05:45.58 - STEP: Creating secret with name s-test-opt-upd-4bbaaf00-5a3b-49af-a40b-16b12aeed372 12/14/22 16:05:45.586 - STEP: Creating the pod 12/14/22 16:05:45.595 - Dec 14 16:05:45.609: INFO: Waiting up to 5m0s for pod "pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281" in namespace "secrets-1497" to be "running and ready" - Dec 14 16:05:45.615: INFO: Pod "pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281": Phase="Pending", Reason="", readiness=false. Elapsed: 6.586148ms - Dec 14 16:05:45.615: INFO: The phase of Pod pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:05:47.623: INFO: Pod "pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01452916s - Dec 14 16:05:47.623: INFO: The phase of Pod pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:05:49.629: INFO: Pod "pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281": Phase="Running", Reason="", readiness=true. Elapsed: 4.020093228s - Dec 14 16:05:49.629: INFO: The phase of Pod pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281 is Running (Ready = true) - Dec 14 16:05:49.629: INFO: Pod "pod-secrets-ea685b3c-a904-4301-94b3-6a63771db281" satisfied condition "running and ready" - STEP: Deleting secret s-test-opt-del-381c6675-cab3-4aff-b10a-0d1be7b17aca 12/14/22 16:05:49.686 - STEP: Updating secret s-test-opt-upd-4bbaaf00-5a3b-49af-a40b-16b12aeed372 12/14/22 16:05:49.698 - STEP: Creating secret with name s-test-opt-create-505b30ff-fe09-4f13-b265-00d433538c8d 12/14/22 16:05:49.706 - STEP: waiting to observe update in volume 12/14/22 16:05:49.715 - [AfterEach] [sig-storage] Secrets + STEP: Creating a kubernetes client 02/25/23 10:10:58.797 + Feb 25 10:10:58.798: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename custom-resource-definition 02/25/23 10:10:58.805 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:10:58.852 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:10:58.858 + [It] should include custom resource definition resources in discovery documents [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:198 + STEP: fetching the /apis discovery document 02/25/23 10:10:58.865 + STEP: finding the apiextensions.k8s.io API group in the /apis discovery document 02/25/23 10:10:58.869 + STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis discovery document 02/25/23 10:10:58.87 + STEP: fetching the /apis/apiextensions.k8s.io discovery document 02/25/23 10:10:58.87 + STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis/apiextensions.k8s.io discovery document 02/25/23 10:10:58.873 + STEP: fetching the /apis/apiextensions.k8s.io/v1 discovery document 02/25/23 10:10:58.874 + STEP: finding customresourcedefinitions resources in the /apis/apiextensions.k8s.io/v1 discovery document 02/25/23 10:10:58.877 + [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:07:02.584: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "secrets-1497" for this suite. 12/14/22 16:07:02.592 + Feb 25 10:10:58.879: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "custom-resource-definition-8987" for this suite. 02/25/23 10:10:58.896 << End Captured GinkgoWriter Output ------------------------------ -SSSSSS +SSSSSSSSSSSSS +------------------------------ +[sig-apps] ControllerRevision [Serial] + should manage the lifecycle of a ControllerRevision [Conformance] + test/e2e/apps/controller_revision.go:124 +[BeforeEach] [sig-apps] ControllerRevision [Serial] + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:10:58.946 +Feb 25 10:10:58.946: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename controllerrevisions 02/25/23 10:10:58.952 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:10:58.986 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:10:58.995 +[BeforeEach] [sig-apps] ControllerRevision [Serial] + test/e2e/apps/controller_revision.go:93 +[It] should manage the lifecycle of a ControllerRevision [Conformance] + test/e2e/apps/controller_revision.go:124 +STEP: Creating DaemonSet "e2e-l27lq-daemon-set" 02/25/23 10:10:59.068 +STEP: Check that daemon pods launch on every node of the cluster. 02/25/23 10:10:59.083 +Feb 25 10:10:59.101: INFO: Number of nodes with available pods controlled by daemonset e2e-l27lq-daemon-set: 0 +Feb 25 10:10:59.101: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 10:11:00.127: INFO: Number of nodes with available pods controlled by daemonset e2e-l27lq-daemon-set: 0 +Feb 25 10:11:00.128: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 10:11:01.197: INFO: Number of nodes with available pods controlled by daemonset e2e-l27lq-daemon-set: 1 +Feb 25 10:11:01.197: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 10:11:02.116: INFO: Number of nodes with available pods controlled by daemonset e2e-l27lq-daemon-set: 2 +Feb 25 10:11:02.116: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 10:11:03.115: INFO: Number of nodes with available pods controlled by daemonset e2e-l27lq-daemon-set: 3 +Feb 25 10:11:03.115: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset e2e-l27lq-daemon-set +STEP: Confirm DaemonSet "e2e-l27lq-daemon-set" successfully created with "daemonset-name=e2e-l27lq-daemon-set" label 02/25/23 10:11:03.119 +STEP: Listing all ControllerRevisions with label "daemonset-name=e2e-l27lq-daemon-set" 02/25/23 10:11:03.136 +Feb 25 10:11:03.145: INFO: Located ControllerRevision: "e2e-l27lq-daemon-set-5498f48cc9" +STEP: Patching ControllerRevision "e2e-l27lq-daemon-set-5498f48cc9" 02/25/23 10:11:03.15 +Feb 25 10:11:03.162: INFO: e2e-l27lq-daemon-set-5498f48cc9 has been patched +STEP: Create a new ControllerRevision 02/25/23 10:11:03.162 +Feb 25 10:11:03.179: INFO: Created ControllerRevision: e2e-l27lq-daemon-set-757b7bb8b7 +STEP: Confirm that there are two ControllerRevisions 02/25/23 10:11:03.179 +Feb 25 10:11:03.180: INFO: Requesting list of ControllerRevisions to confirm quantity +Feb 25 10:11:03.188: INFO: Found 2 ControllerRevisions +STEP: Deleting ControllerRevision "e2e-l27lq-daemon-set-5498f48cc9" 02/25/23 10:11:03.188 +STEP: Confirm that there is only one ControllerRevision 02/25/23 10:11:03.206 +Feb 25 10:11:03.207: INFO: Requesting list of ControllerRevisions to confirm quantity +Feb 25 10:11:03.211: INFO: Found 1 ControllerRevisions +STEP: Updating ControllerRevision "e2e-l27lq-daemon-set-757b7bb8b7" 02/25/23 10:11:03.216 +Feb 25 10:11:03.231: INFO: e2e-l27lq-daemon-set-757b7bb8b7 has been updated +STEP: Generate another ControllerRevision by patching the Daemonset 02/25/23 10:11:03.231 +W0225 10:11:03.245605 13 warnings.go:70] unknown field "updateStrategy" +STEP: Confirm that there are two ControllerRevisions 02/25/23 10:11:03.245 +Feb 25 10:11:03.246: INFO: Requesting list of ControllerRevisions to confirm quantity +Feb 25 10:11:04.253: INFO: Requesting list of ControllerRevisions to confirm quantity +Feb 25 10:11:04.263: INFO: Found 2 ControllerRevisions +STEP: Removing a ControllerRevision via 'DeleteCollection' with labelSelector: "e2e-l27lq-daemon-set-757b7bb8b7=updated" 02/25/23 10:11:04.263 +STEP: Confirm that there is only one ControllerRevision 02/25/23 10:11:04.279 +Feb 25 10:11:04.280: INFO: Requesting list of ControllerRevisions to confirm quantity +Feb 25 10:11:04.287: INFO: Found 1 ControllerRevisions +Feb 25 10:11:04.293: INFO: ControllerRevision "e2e-l27lq-daemon-set-5844b698c4" has revision 3 +[AfterEach] [sig-apps] ControllerRevision [Serial] + test/e2e/apps/controller_revision.go:58 +STEP: Deleting DaemonSet "e2e-l27lq-daemon-set" 02/25/23 10:11:04.302 +STEP: deleting DaemonSet.extensions e2e-l27lq-daemon-set in namespace controllerrevisions-552, will wait for the garbage collector to delete the pods 02/25/23 10:11:04.303 +Feb 25 10:11:04.381: INFO: Deleting DaemonSet.extensions e2e-l27lq-daemon-set took: 18.081088ms +Feb 25 10:11:04.483: INFO: Terminating DaemonSet.extensions e2e-l27lq-daemon-set pods took: 102.039585ms +Feb 25 10:11:05.993: INFO: Number of nodes with available pods controlled by daemonset e2e-l27lq-daemon-set: 0 +Feb 25 10:11:05.994: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset e2e-l27lq-daemon-set +Feb 25 10:11:06.006: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"15673"},"items":null} + +Feb 25 10:11:06.014: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"15673"},"items":null} + +[AfterEach] [sig-apps] ControllerRevision [Serial] + test/e2e/framework/framework.go:187 +Feb 25 10:11:06.096: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "controllerrevisions-552" for this suite. 02/25/23 10:11:06.108 +{"msg":"PASSED [sig-apps] ControllerRevision [Serial] should manage the lifecycle of a ControllerRevision [Conformance]","completed":83,"skipped":1648,"failed":0} +------------------------------ +• [SLOW TEST] [7.175 seconds] +[sig-apps] ControllerRevision [Serial] +test/e2e/apps/framework.go:23 + should manage the lifecycle of a ControllerRevision [Conformance] + test/e2e/apps/controller_revision.go:124 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-apps] ControllerRevision [Serial] + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:10:58.946 + Feb 25 10:10:58.946: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename controllerrevisions 02/25/23 10:10:58.952 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:10:58.986 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:10:58.995 + [BeforeEach] [sig-apps] ControllerRevision [Serial] + test/e2e/apps/controller_revision.go:93 + [It] should manage the lifecycle of a ControllerRevision [Conformance] + test/e2e/apps/controller_revision.go:124 + STEP: Creating DaemonSet "e2e-l27lq-daemon-set" 02/25/23 10:10:59.068 + STEP: Check that daemon pods launch on every node of the cluster. 02/25/23 10:10:59.083 + Feb 25 10:10:59.101: INFO: Number of nodes with available pods controlled by daemonset e2e-l27lq-daemon-set: 0 + Feb 25 10:10:59.101: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 10:11:00.127: INFO: Number of nodes with available pods controlled by daemonset e2e-l27lq-daemon-set: 0 + Feb 25 10:11:00.128: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 10:11:01.197: INFO: Number of nodes with available pods controlled by daemonset e2e-l27lq-daemon-set: 1 + Feb 25 10:11:01.197: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 10:11:02.116: INFO: Number of nodes with available pods controlled by daemonset e2e-l27lq-daemon-set: 2 + Feb 25 10:11:02.116: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 10:11:03.115: INFO: Number of nodes with available pods controlled by daemonset e2e-l27lq-daemon-set: 3 + Feb 25 10:11:03.115: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset e2e-l27lq-daemon-set + STEP: Confirm DaemonSet "e2e-l27lq-daemon-set" successfully created with "daemonset-name=e2e-l27lq-daemon-set" label 02/25/23 10:11:03.119 + STEP: Listing all ControllerRevisions with label "daemonset-name=e2e-l27lq-daemon-set" 02/25/23 10:11:03.136 + Feb 25 10:11:03.145: INFO: Located ControllerRevision: "e2e-l27lq-daemon-set-5498f48cc9" + STEP: Patching ControllerRevision "e2e-l27lq-daemon-set-5498f48cc9" 02/25/23 10:11:03.15 + Feb 25 10:11:03.162: INFO: e2e-l27lq-daemon-set-5498f48cc9 has been patched + STEP: Create a new ControllerRevision 02/25/23 10:11:03.162 + Feb 25 10:11:03.179: INFO: Created ControllerRevision: e2e-l27lq-daemon-set-757b7bb8b7 + STEP: Confirm that there are two ControllerRevisions 02/25/23 10:11:03.179 + Feb 25 10:11:03.180: INFO: Requesting list of ControllerRevisions to confirm quantity + Feb 25 10:11:03.188: INFO: Found 2 ControllerRevisions + STEP: Deleting ControllerRevision "e2e-l27lq-daemon-set-5498f48cc9" 02/25/23 10:11:03.188 + STEP: Confirm that there is only one ControllerRevision 02/25/23 10:11:03.206 + Feb 25 10:11:03.207: INFO: Requesting list of ControllerRevisions to confirm quantity + Feb 25 10:11:03.211: INFO: Found 1 ControllerRevisions + STEP: Updating ControllerRevision "e2e-l27lq-daemon-set-757b7bb8b7" 02/25/23 10:11:03.216 + Feb 25 10:11:03.231: INFO: e2e-l27lq-daemon-set-757b7bb8b7 has been updated + STEP: Generate another ControllerRevision by patching the Daemonset 02/25/23 10:11:03.231 + W0225 10:11:03.245605 13 warnings.go:70] unknown field "updateStrategy" + STEP: Confirm that there are two ControllerRevisions 02/25/23 10:11:03.245 + Feb 25 10:11:03.246: INFO: Requesting list of ControllerRevisions to confirm quantity + Feb 25 10:11:04.253: INFO: Requesting list of ControllerRevisions to confirm quantity + Feb 25 10:11:04.263: INFO: Found 2 ControllerRevisions + STEP: Removing a ControllerRevision via 'DeleteCollection' with labelSelector: "e2e-l27lq-daemon-set-757b7bb8b7=updated" 02/25/23 10:11:04.263 + STEP: Confirm that there is only one ControllerRevision 02/25/23 10:11:04.279 + Feb 25 10:11:04.280: INFO: Requesting list of ControllerRevisions to confirm quantity + Feb 25 10:11:04.287: INFO: Found 1 ControllerRevisions + Feb 25 10:11:04.293: INFO: ControllerRevision "e2e-l27lq-daemon-set-5844b698c4" has revision 3 + [AfterEach] [sig-apps] ControllerRevision [Serial] + test/e2e/apps/controller_revision.go:58 + STEP: Deleting DaemonSet "e2e-l27lq-daemon-set" 02/25/23 10:11:04.302 + STEP: deleting DaemonSet.extensions e2e-l27lq-daemon-set in namespace controllerrevisions-552, will wait for the garbage collector to delete the pods 02/25/23 10:11:04.303 + Feb 25 10:11:04.381: INFO: Deleting DaemonSet.extensions e2e-l27lq-daemon-set took: 18.081088ms + Feb 25 10:11:04.483: INFO: Terminating DaemonSet.extensions e2e-l27lq-daemon-set pods took: 102.039585ms + Feb 25 10:11:05.993: INFO: Number of nodes with available pods controlled by daemonset e2e-l27lq-daemon-set: 0 + Feb 25 10:11:05.994: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset e2e-l27lq-daemon-set + Feb 25 10:11:06.006: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"15673"},"items":null} + + Feb 25 10:11:06.014: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"15673"},"items":null} + + [AfterEach] [sig-apps] ControllerRevision [Serial] + test/e2e/framework/framework.go:187 + Feb 25 10:11:06.096: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "controllerrevisions-552" for this suite. 02/25/23 10:11:06.108 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] Pods - should get a host IP [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:203 + should run through the lifecycle of Pods and PodStatus [Conformance] + test/e2e/common/node/pods.go:895 [BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:07:02.606 -Dec 14 16:07:02.606: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename pods 12/14/22 16:07:02.611 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:02.647 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:02.651 +STEP: Creating a kubernetes client 02/25/23 10:11:06.196 +Feb 25 10:11:06.197: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename pods 02/25/23 10:11:06.201 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:06.241 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:06.25 [BeforeEach] [sig-node] Pods test/e2e/common/node/pods.go:193 -[It] should get a host IP [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:203 -STEP: creating pod 12/14/22 16:07:02.656 -Dec 14 16:07:02.670: INFO: Waiting up to 5m0s for pod "pod-hostip-7e526d1c-d89a-4ad8-8adb-7cd89971529e" in namespace "pods-9947" to be "running and ready" -Dec 14 16:07:02.676: INFO: Pod "pod-hostip-7e526d1c-d89a-4ad8-8adb-7cd89971529e": Phase="Pending", Reason="", readiness=false. Elapsed: 6.50608ms -Dec 14 16:07:02.677: INFO: The phase of Pod pod-hostip-7e526d1c-d89a-4ad8-8adb-7cd89971529e is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:07:04.686: INFO: Pod "pod-hostip-7e526d1c-d89a-4ad8-8adb-7cd89971529e": Phase="Running", Reason="", readiness=true. Elapsed: 2.015905773s -Dec 14 16:07:04.686: INFO: The phase of Pod pod-hostip-7e526d1c-d89a-4ad8-8adb-7cd89971529e is Running (Ready = true) -Dec 14 16:07:04.686: INFO: Pod "pod-hostip-7e526d1c-d89a-4ad8-8adb-7cd89971529e" satisfied condition "running and ready" -Dec 14 16:07:04.700: INFO: Pod pod-hostip-7e526d1c-d89a-4ad8-8adb-7cd89971529e has hostIP: 192.168.121.249 +[It] should run through the lifecycle of Pods and PodStatus [Conformance] + test/e2e/common/node/pods.go:895 +STEP: creating a Pod with a static label 02/25/23 10:11:06.273 +STEP: watching for Pod to be ready 02/25/23 10:11:06.292 +Feb 25 10:11:06.294: INFO: observed Pod pod-test in namespace pods-8232 in phase Pending with labels: map[test-pod-static:true] & conditions [] +Feb 25 10:11:06.316: INFO: observed Pod pod-test in namespace pods-8232 in phase Pending with labels: map[test-pod-static:true] & conditions [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:06 +0000 UTC }] +Feb 25 10:11:06.342: INFO: observed Pod pod-test in namespace pods-8232 in phase Pending with labels: map[test-pod-static:true] & conditions [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:06 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:06 +0000 UTC ContainersNotReady containers with unready status: [pod-test]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:06 +0000 UTC ContainersNotReady containers with unready status: [pod-test]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:06 +0000 UTC }] +Feb 25 10:11:08.230: INFO: Found Pod pod-test in namespace pods-8232 in phase Running with labels: map[test-pod-static:true] & conditions [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:06 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:08 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:08 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:06 +0000 UTC }] +STEP: patching the Pod with a new Label and updated data 02/25/23 10:11:08.24 +STEP: getting the Pod and ensuring that it's patched 02/25/23 10:11:08.265 +STEP: replacing the Pod's status Ready condition to False 02/25/23 10:11:08.276 +STEP: check the Pod again to ensure its Ready conditions are False 02/25/23 10:11:08.32 +STEP: deleting the Pod via a Collection with a LabelSelector 02/25/23 10:11:08.32 +STEP: watching for the Pod to be deleted 02/25/23 10:11:08.35 +Feb 25 10:11:08.355: INFO: observed event type MODIFIED +Feb 25 10:11:10.269: INFO: observed event type MODIFIED +Feb 25 10:11:11.273: INFO: observed event type MODIFIED +Feb 25 10:11:11.304: INFO: observed event type MODIFIED [AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 -Dec 14 16:07:04.700: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-9947" for this suite. 12/14/22 16:07:04.709 -{"msg":"PASSED [sig-node] Pods should get a host IP [NodeConformance] [Conformance]","completed":82,"skipped":1504,"failed":0} +Feb 25 10:11:11.363: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-8232" for this suite. 02/25/23 10:11:11.375 +{"msg":"PASSED [sig-node] Pods should run through the lifecycle of Pods and PodStatus [Conformance]","completed":84,"skipped":1761,"failed":0} ------------------------------ -• [2.113 seconds] +• [SLOW TEST] [5.195 seconds] [sig-node] Pods test/e2e/common/node/framework.go:23 - should get a host IP [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:203 + should run through the lifecycle of Pods and PodStatus [Conformance] + test/e2e/common/node/pods.go:895 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:07:02.606 - Dec 14 16:07:02.606: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename pods 12/14/22 16:07:02.611 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:02.647 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:02.651 + STEP: Creating a kubernetes client 02/25/23 10:11:06.196 + Feb 25 10:11:06.197: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename pods 02/25/23 10:11:06.201 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:06.241 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:06.25 [BeforeEach] [sig-node] Pods test/e2e/common/node/pods.go:193 - [It] should get a host IP [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:203 - STEP: creating pod 12/14/22 16:07:02.656 - Dec 14 16:07:02.670: INFO: Waiting up to 5m0s for pod "pod-hostip-7e526d1c-d89a-4ad8-8adb-7cd89971529e" in namespace "pods-9947" to be "running and ready" - Dec 14 16:07:02.676: INFO: Pod "pod-hostip-7e526d1c-d89a-4ad8-8adb-7cd89971529e": Phase="Pending", Reason="", readiness=false. Elapsed: 6.50608ms - Dec 14 16:07:02.677: INFO: The phase of Pod pod-hostip-7e526d1c-d89a-4ad8-8adb-7cd89971529e is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:07:04.686: INFO: Pod "pod-hostip-7e526d1c-d89a-4ad8-8adb-7cd89971529e": Phase="Running", Reason="", readiness=true. Elapsed: 2.015905773s - Dec 14 16:07:04.686: INFO: The phase of Pod pod-hostip-7e526d1c-d89a-4ad8-8adb-7cd89971529e is Running (Ready = true) - Dec 14 16:07:04.686: INFO: Pod "pod-hostip-7e526d1c-d89a-4ad8-8adb-7cd89971529e" satisfied condition "running and ready" - Dec 14 16:07:04.700: INFO: Pod pod-hostip-7e526d1c-d89a-4ad8-8adb-7cd89971529e has hostIP: 192.168.121.249 + [It] should run through the lifecycle of Pods and PodStatus [Conformance] + test/e2e/common/node/pods.go:895 + STEP: creating a Pod with a static label 02/25/23 10:11:06.273 + STEP: watching for Pod to be ready 02/25/23 10:11:06.292 + Feb 25 10:11:06.294: INFO: observed Pod pod-test in namespace pods-8232 in phase Pending with labels: map[test-pod-static:true] & conditions [] + Feb 25 10:11:06.316: INFO: observed Pod pod-test in namespace pods-8232 in phase Pending with labels: map[test-pod-static:true] & conditions [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:06 +0000 UTC }] + Feb 25 10:11:06.342: INFO: observed Pod pod-test in namespace pods-8232 in phase Pending with labels: map[test-pod-static:true] & conditions [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:06 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:06 +0000 UTC ContainersNotReady containers with unready status: [pod-test]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:06 +0000 UTC ContainersNotReady containers with unready status: [pod-test]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:06 +0000 UTC }] + Feb 25 10:11:08.230: INFO: Found Pod pod-test in namespace pods-8232 in phase Running with labels: map[test-pod-static:true] & conditions [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:06 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:08 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:08 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 10:11:06 +0000 UTC }] + STEP: patching the Pod with a new Label and updated data 02/25/23 10:11:08.24 + STEP: getting the Pod and ensuring that it's patched 02/25/23 10:11:08.265 + STEP: replacing the Pod's status Ready condition to False 02/25/23 10:11:08.276 + STEP: check the Pod again to ensure its Ready conditions are False 02/25/23 10:11:08.32 + STEP: deleting the Pod via a Collection with a LabelSelector 02/25/23 10:11:08.32 + STEP: watching for the Pod to be deleted 02/25/23 10:11:08.35 + Feb 25 10:11:08.355: INFO: observed event type MODIFIED + Feb 25 10:11:10.269: INFO: observed event type MODIFIED + Feb 25 10:11:11.273: INFO: observed event type MODIFIED + Feb 25 10:11:11.304: INFO: observed event type MODIFIED [AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 - Dec 14 16:07:04.700: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "pods-9947" for this suite. 12/14/22 16:07:04.709 + Feb 25 10:11:11.363: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "pods-8232" for this suite. 02/25/23 10:11:11.375 << End Captured GinkgoWriter Output ------------------------------ -SSSS +SSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-apps] Job - should manage the lifecycle of a job [Conformance] - test/e2e/apps/job.go:531 -[BeforeEach] [sig-apps] Job +[sig-storage] Projected combined + should project all components that make up the projection API [Projection][NodeConformance] [Conformance] + test/e2e/common/storage/projected_combined.go:43 +[BeforeEach] [sig-storage] Projected combined test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:07:04.72 -Dec 14 16:07:04.720: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename job 12/14/22 16:07:04.724 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:04.758 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:04.761 -[It] should manage the lifecycle of a job [Conformance] - test/e2e/apps/job.go:531 -STEP: Creating a suspended job 12/14/22 16:07:04.772 -STEP: Patching the Job 12/14/22 16:07:04.784 -STEP: Watching for Job to be patched 12/14/22 16:07:04.807 -Dec 14 16:07:04.812: INFO: Event ADDED observed for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7] and annotations: map[batch.kubernetes.io/job-tracking:] -Dec 14 16:07:04.812: INFO: Event MODIFIED observed for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7] and annotations: map[batch.kubernetes.io/job-tracking:] -Dec 14 16:07:04.812: INFO: Event MODIFIED found for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking:] -STEP: Updating the job 12/14/22 16:07:04.812 -STEP: Watching for Job to be updated 12/14/22 16:07:04.827 -Dec 14 16:07:04.829: INFO: Event MODIFIED found for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] -Dec 14 16:07:04.829: INFO: Found Job annotations: map[string]string{"batch.kubernetes.io/job-tracking":"", "updated":"true"} -STEP: Listing all Jobs with LabelSelector 12/14/22 16:07:04.83 -Dec 14 16:07:04.834: INFO: Job: e2e-rqtm7 as labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] -STEP: Waiting for job to complete 12/14/22 16:07:04.834 -STEP: Delete a job collection with a labelselector 12/14/22 16:07:14.843 -STEP: Watching for Job to be deleted 12/14/22 16:07:14.857 -Dec 14 16:07:14.860: INFO: Event MODIFIED observed for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] -Dec 14 16:07:14.860: INFO: Event MODIFIED observed for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] -Dec 14 16:07:14.861: INFO: Event MODIFIED observed for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] -Dec 14 16:07:14.861: INFO: Event MODIFIED observed for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] -Dec 14 16:07:14.861: INFO: Event MODIFIED observed for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] -Dec 14 16:07:14.862: INFO: Event DELETED found for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] -STEP: Relist jobs to confirm deletion 12/14/22 16:07:14.862 -[AfterEach] [sig-apps] Job +STEP: Creating a kubernetes client 02/25/23 10:11:11.42 +Feb 25 10:11:11.421: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 10:11:11.427 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:11.471 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:11.48 +[It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] + test/e2e/common/storage/projected_combined.go:43 +STEP: Creating configMap with name configmap-projected-all-test-volume-91410dd6-0629-48b1-ae01-aec59ce35546 02/25/23 10:11:11.488 +STEP: Creating secret with name secret-projected-all-test-volume-a299c969-e84f-42e7-b67e-5a0b8e8ab5ad 02/25/23 10:11:11.5 +STEP: Creating a pod to test Check all projections for projected volume plugin 02/25/23 10:11:11.513 +Feb 25 10:11:11.538: INFO: Waiting up to 5m0s for pod "projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6" in namespace "projected-6632" to be "Succeeded or Failed" +Feb 25 10:11:11.552: INFO: Pod "projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6": Phase="Pending", Reason="", readiness=false. Elapsed: 13.685609ms +Feb 25 10:11:13.560: INFO: Pod "projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021896316s +Feb 25 10:11:15.562: INFO: Pod "projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.02384769s +STEP: Saw pod success 02/25/23 10:11:15.562 +Feb 25 10:11:15.563: INFO: Pod "projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6" satisfied condition "Succeeded or Failed" +Feb 25 10:11:15.572: INFO: Trying to get logs from node diek3dua9gay-3 pod projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6 container projected-all-volume-test: +STEP: delete the pod 02/25/23 10:11:15.601 +Feb 25 10:11:15.632: INFO: Waiting for pod projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6 to disappear +Feb 25 10:11:15.641: INFO: Pod projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6 no longer exists +[AfterEach] [sig-storage] Projected combined test/e2e/framework/framework.go:187 -Dec 14 16:07:14.868: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "job-3851" for this suite. 12/14/22 16:07:14.879 -{"msg":"PASSED [sig-apps] Job should manage the lifecycle of a job [Conformance]","completed":83,"skipped":1508,"failed":0} +Feb 25 10:11:15.641: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-6632" for this suite. 02/25/23 10:11:15.651 +{"msg":"PASSED [sig-storage] Projected combined should project all components that make up the projection API [Projection][NodeConformance] [Conformance]","completed":85,"skipped":1787,"failed":0} ------------------------------ -• [SLOW TEST] [10.169 seconds] -[sig-apps] Job -test/e2e/apps/framework.go:23 - should manage the lifecycle of a job [Conformance] - test/e2e/apps/job.go:531 +• [4.250 seconds] +[sig-storage] Projected combined +test/e2e/common/storage/framework.go:23 + should project all components that make up the projection API [Projection][NodeConformance] [Conformance] + test/e2e/common/storage/projected_combined.go:43 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Job + [BeforeEach] [sig-storage] Projected combined test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:07:04.72 - Dec 14 16:07:04.720: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename job 12/14/22 16:07:04.724 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:04.758 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:04.761 - [It] should manage the lifecycle of a job [Conformance] - test/e2e/apps/job.go:531 - STEP: Creating a suspended job 12/14/22 16:07:04.772 - STEP: Patching the Job 12/14/22 16:07:04.784 - STEP: Watching for Job to be patched 12/14/22 16:07:04.807 - Dec 14 16:07:04.812: INFO: Event ADDED observed for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7] and annotations: map[batch.kubernetes.io/job-tracking:] - Dec 14 16:07:04.812: INFO: Event MODIFIED observed for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7] and annotations: map[batch.kubernetes.io/job-tracking:] - Dec 14 16:07:04.812: INFO: Event MODIFIED found for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking:] - STEP: Updating the job 12/14/22 16:07:04.812 - STEP: Watching for Job to be updated 12/14/22 16:07:04.827 - Dec 14 16:07:04.829: INFO: Event MODIFIED found for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] - Dec 14 16:07:04.829: INFO: Found Job annotations: map[string]string{"batch.kubernetes.io/job-tracking":"", "updated":"true"} - STEP: Listing all Jobs with LabelSelector 12/14/22 16:07:04.83 - Dec 14 16:07:04.834: INFO: Job: e2e-rqtm7 as labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] - STEP: Waiting for job to complete 12/14/22 16:07:04.834 - STEP: Delete a job collection with a labelselector 12/14/22 16:07:14.843 - STEP: Watching for Job to be deleted 12/14/22 16:07:14.857 - Dec 14 16:07:14.860: INFO: Event MODIFIED observed for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] - Dec 14 16:07:14.860: INFO: Event MODIFIED observed for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] - Dec 14 16:07:14.861: INFO: Event MODIFIED observed for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] - Dec 14 16:07:14.861: INFO: Event MODIFIED observed for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] - Dec 14 16:07:14.861: INFO: Event MODIFIED observed for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] - Dec 14 16:07:14.862: INFO: Event DELETED found for Job e2e-rqtm7 in namespace job-3851 with labels: map[e2e-job-label:e2e-rqtm7 e2e-rqtm7:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] - STEP: Relist jobs to confirm deletion 12/14/22 16:07:14.862 - [AfterEach] [sig-apps] Job + STEP: Creating a kubernetes client 02/25/23 10:11:11.42 + Feb 25 10:11:11.421: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 10:11:11.427 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:11.471 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:11.48 + [It] should project all components that make up the projection API [Projection][NodeConformance] [Conformance] + test/e2e/common/storage/projected_combined.go:43 + STEP: Creating configMap with name configmap-projected-all-test-volume-91410dd6-0629-48b1-ae01-aec59ce35546 02/25/23 10:11:11.488 + STEP: Creating secret with name secret-projected-all-test-volume-a299c969-e84f-42e7-b67e-5a0b8e8ab5ad 02/25/23 10:11:11.5 + STEP: Creating a pod to test Check all projections for projected volume plugin 02/25/23 10:11:11.513 + Feb 25 10:11:11.538: INFO: Waiting up to 5m0s for pod "projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6" in namespace "projected-6632" to be "Succeeded or Failed" + Feb 25 10:11:11.552: INFO: Pod "projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6": Phase="Pending", Reason="", readiness=false. Elapsed: 13.685609ms + Feb 25 10:11:13.560: INFO: Pod "projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021896316s + Feb 25 10:11:15.562: INFO: Pod "projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.02384769s + STEP: Saw pod success 02/25/23 10:11:15.562 + Feb 25 10:11:15.563: INFO: Pod "projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6" satisfied condition "Succeeded or Failed" + Feb 25 10:11:15.572: INFO: Trying to get logs from node diek3dua9gay-3 pod projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6 container projected-all-volume-test: + STEP: delete the pod 02/25/23 10:11:15.601 + Feb 25 10:11:15.632: INFO: Waiting for pod projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6 to disappear + Feb 25 10:11:15.641: INFO: Pod projected-volume-cd76176b-c832-46ef-8eea-2be949575fd6 no longer exists + [AfterEach] [sig-storage] Projected combined test/e2e/framework/framework.go:187 - Dec 14 16:07:14.868: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "job-3851" for this suite. 12/14/22 16:07:14.879 + Feb 25 10:11:15.641: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-6632" for this suite. 02/25/23 10:11:15.651 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSS +S ------------------------------ -[sig-node] Container Runtime blackbox test on terminated container - should report termination message from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:215 -[BeforeEach] [sig-node] Container Runtime +[sig-storage] Projected downwardAPI + should update annotations on modification [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:161 +[BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:07:14.89 -Dec 14 16:07:14.891: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-runtime 12/14/22 16:07:14.892 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:14.984 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:14.989 -[It] should report termination message from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:215 -STEP: create the container 12/14/22 16:07:14.993 -STEP: wait for the container to reach Failed 12/14/22 16:07:15.001 -STEP: get the container status 12/14/22 16:07:18.026 -STEP: the container should be terminated 12/14/22 16:07:18.029 -STEP: the termination message should be set 12/14/22 16:07:18.03 -Dec 14 16:07:18.030: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- -STEP: delete the container 12/14/22 16:07:18.03 -[AfterEach] [sig-node] Container Runtime +STEP: Creating a kubernetes client 02/25/23 10:11:15.67 +Feb 25 10:11:15.671: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 10:11:15.675 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:15.722 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:15.726 +[BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 +[It] should update annotations on modification [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:161 +STEP: Creating the pod 02/25/23 10:11:15.732 +Feb 25 10:11:15.753: INFO: Waiting up to 5m0s for pod "annotationupdatee06f1b17-539a-437a-900d-177f366f889f" in namespace "projected-4053" to be "running and ready" +Feb 25 10:11:15.759: INFO: Pod "annotationupdatee06f1b17-539a-437a-900d-177f366f889f": Phase="Pending", Reason="", readiness=false. Elapsed: 6.092483ms +Feb 25 10:11:15.759: INFO: The phase of Pod annotationupdatee06f1b17-539a-437a-900d-177f366f889f is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:11:17.770: INFO: Pod "annotationupdatee06f1b17-539a-437a-900d-177f366f889f": Phase="Running", Reason="", readiness=true. Elapsed: 2.016764045s +Feb 25 10:11:17.770: INFO: The phase of Pod annotationupdatee06f1b17-539a-437a-900d-177f366f889f is Running (Ready = true) +Feb 25 10:11:17.770: INFO: Pod "annotationupdatee06f1b17-539a-437a-900d-177f366f889f" satisfied condition "running and ready" +Feb 25 10:11:18.325: INFO: Successfully updated pod "annotationupdatee06f1b17-539a-437a-900d-177f366f889f" +[AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 -Dec 14 16:07:18.049: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-runtime-5315" for this suite. 12/14/22 16:07:18.054 -{"msg":"PASSED [sig-node] Container Runtime blackbox test on terminated container should report termination message from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","completed":84,"skipped":1515,"failed":0} +Feb 25 10:11:20.359: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-4053" for this suite. 02/25/23 10:11:20.368 +{"msg":"PASSED [sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance]","completed":86,"skipped":1788,"failed":0} ------------------------------ -• [3.175 seconds] -[sig-node] Container Runtime -test/e2e/common/node/framework.go:23 - blackbox test - test/e2e/common/node/runtime.go:43 - on terminated container - test/e2e/common/node/runtime.go:136 - should report termination message from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:215 +• [4.713 seconds] +[sig-storage] Projected downwardAPI +test/e2e/common/storage/framework.go:23 + should update annotations on modification [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:161 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Container Runtime + [BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:07:14.89 - Dec 14 16:07:14.891: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-runtime 12/14/22 16:07:14.892 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:14.984 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:14.989 - [It] should report termination message from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:215 - STEP: create the container 12/14/22 16:07:14.993 - STEP: wait for the container to reach Failed 12/14/22 16:07:15.001 - STEP: get the container status 12/14/22 16:07:18.026 - STEP: the container should be terminated 12/14/22 16:07:18.029 - STEP: the termination message should be set 12/14/22 16:07:18.03 - Dec 14 16:07:18.030: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- - STEP: delete the container 12/14/22 16:07:18.03 - [AfterEach] [sig-node] Container Runtime + STEP: Creating a kubernetes client 02/25/23 10:11:15.67 + Feb 25 10:11:15.671: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 10:11:15.675 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:15.722 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:15.726 + [BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 + [It] should update annotations on modification [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:161 + STEP: Creating the pod 02/25/23 10:11:15.732 + Feb 25 10:11:15.753: INFO: Waiting up to 5m0s for pod "annotationupdatee06f1b17-539a-437a-900d-177f366f889f" in namespace "projected-4053" to be "running and ready" + Feb 25 10:11:15.759: INFO: Pod "annotationupdatee06f1b17-539a-437a-900d-177f366f889f": Phase="Pending", Reason="", readiness=false. Elapsed: 6.092483ms + Feb 25 10:11:15.759: INFO: The phase of Pod annotationupdatee06f1b17-539a-437a-900d-177f366f889f is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:11:17.770: INFO: Pod "annotationupdatee06f1b17-539a-437a-900d-177f366f889f": Phase="Running", Reason="", readiness=true. Elapsed: 2.016764045s + Feb 25 10:11:17.770: INFO: The phase of Pod annotationupdatee06f1b17-539a-437a-900d-177f366f889f is Running (Ready = true) + Feb 25 10:11:17.770: INFO: Pod "annotationupdatee06f1b17-539a-437a-900d-177f366f889f" satisfied condition "running and ready" + Feb 25 10:11:18.325: INFO: Successfully updated pod "annotationupdatee06f1b17-539a-437a-900d-177f366f889f" + [AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 - Dec 14 16:07:18.049: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-runtime-5315" for this suite. 12/14/22 16:07:18.054 + Feb 25 10:11:20.359: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-4053" for this suite. 02/25/23 10:11:20.368 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSS +SSSSSSSSSSSS ------------------------------ -[sig-node] Kubelet when scheduling a busybox command that always fails in a pod - should be possible to delete [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:135 -[BeforeEach] [sig-node] Kubelet +[sig-node] Secrets + should be consumable from pods in env vars [NodeConformance] [Conformance] + test/e2e/common/node/secrets.go:45 +[BeforeEach] [sig-node] Secrets test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:07:18.074 -Dec 14 16:07:18.074: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubelet-test 12/14/22 16:07:18.077 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:18.097 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:18.1 -[BeforeEach] [sig-node] Kubelet - test/e2e/common/node/kubelet.go:41 -[BeforeEach] when scheduling a busybox command that always fails in a pod - test/e2e/common/node/kubelet.go:85 -[It] should be possible to delete [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:135 -[AfterEach] [sig-node] Kubelet +STEP: Creating a kubernetes client 02/25/23 10:11:20.385 +Feb 25 10:11:20.385: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename secrets 02/25/23 10:11:20.388 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:20.435 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:20.44 +[It] should be consumable from pods in env vars [NodeConformance] [Conformance] + test/e2e/common/node/secrets.go:45 +STEP: Creating secret with name secret-test-a4783d95-9e80-4fcf-aac3-402d05e24251 02/25/23 10:11:20.446 +STEP: Creating a pod to test consume secrets 02/25/23 10:11:20.458 +Feb 25 10:11:20.477: INFO: Waiting up to 5m0s for pod "pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90" in namespace "secrets-1623" to be "Succeeded or Failed" +Feb 25 10:11:20.484: INFO: Pod "pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90": Phase="Pending", Reason="", readiness=false. Elapsed: 6.11821ms +Feb 25 10:11:22.494: INFO: Pod "pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016940611s +Feb 25 10:11:24.493: INFO: Pod "pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90": Phase="Running", Reason="", readiness=true. Elapsed: 4.015105722s +Feb 25 10:11:26.493: INFO: Pod "pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90": Phase="Running", Reason="", readiness=false. Elapsed: 6.01530305s +Feb 25 10:11:28.495: INFO: Pod "pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.017820306s +STEP: Saw pod success 02/25/23 10:11:28.495 +Feb 25 10:11:28.496: INFO: Pod "pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90" satisfied condition "Succeeded or Failed" +Feb 25 10:11:28.505: INFO: Trying to get logs from node diek3dua9gay-2 pod pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90 container secret-env-test: +STEP: delete the pod 02/25/23 10:11:28.535 +Feb 25 10:11:28.562: INFO: Waiting for pod pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90 to disappear +Feb 25 10:11:28.576: INFO: Pod pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90 no longer exists +[AfterEach] [sig-node] Secrets test/e2e/framework/framework.go:187 -Dec 14 16:07:18.149: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubelet-test-1433" for this suite. 12/14/22 16:07:18.155 -{"msg":"PASSED [sig-node] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance]","completed":85,"skipped":1531,"failed":0} +Feb 25 10:11:28.577: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-1623" for this suite. 02/25/23 10:11:28.596 +{"msg":"PASSED [sig-node] Secrets should be consumable from pods in env vars [NodeConformance] [Conformance]","completed":87,"skipped":1800,"failed":0} ------------------------------ -• [0.101 seconds] -[sig-node] Kubelet +• [SLOW TEST] [8.227 seconds] +[sig-node] Secrets test/e2e/common/node/framework.go:23 - when scheduling a busybox command that always fails in a pod - test/e2e/common/node/kubelet.go:82 - should be possible to delete [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:135 + should be consumable from pods in env vars [NodeConformance] [Conformance] + test/e2e/common/node/secrets.go:45 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Kubelet + [BeforeEach] [sig-node] Secrets test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:07:18.074 - Dec 14 16:07:18.074: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubelet-test 12/14/22 16:07:18.077 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:18.097 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:18.1 - [BeforeEach] [sig-node] Kubelet - test/e2e/common/node/kubelet.go:41 - [BeforeEach] when scheduling a busybox command that always fails in a pod - test/e2e/common/node/kubelet.go:85 - [It] should be possible to delete [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:135 - [AfterEach] [sig-node] Kubelet + STEP: Creating a kubernetes client 02/25/23 10:11:20.385 + Feb 25 10:11:20.385: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename secrets 02/25/23 10:11:20.388 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:20.435 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:20.44 + [It] should be consumable from pods in env vars [NodeConformance] [Conformance] + test/e2e/common/node/secrets.go:45 + STEP: Creating secret with name secret-test-a4783d95-9e80-4fcf-aac3-402d05e24251 02/25/23 10:11:20.446 + STEP: Creating a pod to test consume secrets 02/25/23 10:11:20.458 + Feb 25 10:11:20.477: INFO: Waiting up to 5m0s for pod "pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90" in namespace "secrets-1623" to be "Succeeded or Failed" + Feb 25 10:11:20.484: INFO: Pod "pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90": Phase="Pending", Reason="", readiness=false. Elapsed: 6.11821ms + Feb 25 10:11:22.494: INFO: Pod "pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016940611s + Feb 25 10:11:24.493: INFO: Pod "pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90": Phase="Running", Reason="", readiness=true. Elapsed: 4.015105722s + Feb 25 10:11:26.493: INFO: Pod "pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90": Phase="Running", Reason="", readiness=false. Elapsed: 6.01530305s + Feb 25 10:11:28.495: INFO: Pod "pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.017820306s + STEP: Saw pod success 02/25/23 10:11:28.495 + Feb 25 10:11:28.496: INFO: Pod "pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90" satisfied condition "Succeeded or Failed" + Feb 25 10:11:28.505: INFO: Trying to get logs from node diek3dua9gay-2 pod pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90 container secret-env-test: + STEP: delete the pod 02/25/23 10:11:28.535 + Feb 25 10:11:28.562: INFO: Waiting for pod pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90 to disappear + Feb 25 10:11:28.576: INFO: Pod pod-secrets-38ba794b-c64d-4276-969f-b48a51d84f90 no longer exists + [AfterEach] [sig-node] Secrets test/e2e/framework/framework.go:187 - Dec 14 16:07:18.149: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubelet-test-1433" for this suite. 12/14/22 16:07:18.155 + Feb 25 10:11:28.577: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "secrets-1623" for this suite. 02/25/23 10:11:28.596 << End Captured GinkgoWriter Output ------------------------------ -SS +SSSSSSSSSSSSSSSSSS ------------------------------ -[sig-auth] ServiceAccounts - should run through the lifecycle of a ServiceAccount [Conformance] - test/e2e/auth/service_accounts.go:646 -[BeforeEach] [sig-auth] ServiceAccounts +[sig-apps] ReplicationController + should serve a basic image on each replica with a public image [Conformance] + test/e2e/apps/rc.go:66 +[BeforeEach] [sig-apps] ReplicationController test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:07:18.178 -Dec 14 16:07:18.178: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename svcaccounts 12/14/22 16:07:18.18 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:18.203 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:18.207 -[It] should run through the lifecycle of a ServiceAccount [Conformance] - test/e2e/auth/service_accounts.go:646 -STEP: creating a ServiceAccount 12/14/22 16:07:18.21 -STEP: watching for the ServiceAccount to be added 12/14/22 16:07:18.219 -STEP: patching the ServiceAccount 12/14/22 16:07:18.221 -STEP: finding ServiceAccount in list of all ServiceAccounts (by LabelSelector) 12/14/22 16:07:18.231 -STEP: deleting the ServiceAccount 12/14/22 16:07:18.235 -[AfterEach] [sig-auth] ServiceAccounts +STEP: Creating a kubernetes client 02/25/23 10:11:28.629 +Feb 25 10:11:28.629: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename replication-controller 02/25/23 10:11:28.631 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:28.67 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:28.68 +[BeforeEach] [sig-apps] ReplicationController + test/e2e/apps/rc.go:56 +[It] should serve a basic image on each replica with a public image [Conformance] + test/e2e/apps/rc.go:66 +STEP: Creating replication controller my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f 02/25/23 10:11:28.686 +Feb 25 10:11:28.705: INFO: Pod name my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f: Found 0 pods out of 1 +Feb 25 10:11:33.717: INFO: Pod name my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f: Found 1 pods out of 1 +Feb 25 10:11:33.717: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f" are running +Feb 25 10:11:33.717: INFO: Waiting up to 5m0s for pod "my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f-ld2bv" in namespace "replication-controller-5303" to be "running" +Feb 25 10:11:33.728: INFO: Pod "my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f-ld2bv": Phase="Running", Reason="", readiness=true. Elapsed: 10.776153ms +Feb 25 10:11:33.728: INFO: Pod "my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f-ld2bv" satisfied condition "running" +Feb 25 10:11:33.728: INFO: Pod "my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f-ld2bv" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:11:28 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:11:30 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:11:30 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:11:28 +0000 UTC Reason: Message:}]) +Feb 25 10:11:33.728: INFO: Trying to dial the pod +Feb 25 10:11:38.763: INFO: Controller my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f: Got expected result from replica 1 [my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f-ld2bv]: "my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f-ld2bv", 1 of 1 required successes so far +[AfterEach] [sig-apps] ReplicationController test/e2e/framework/framework.go:187 -Dec 14 16:07:18.254: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "svcaccounts-9374" for this suite. 12/14/22 16:07:18.259 -{"msg":"PASSED [sig-auth] ServiceAccounts should run through the lifecycle of a ServiceAccount [Conformance]","completed":86,"skipped":1533,"failed":0} +Feb 25 10:11:38.764: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "replication-controller-5303" for this suite. 02/25/23 10:11:38.775 +{"msg":"PASSED [sig-apps] ReplicationController should serve a basic image on each replica with a public image [Conformance]","completed":88,"skipped":1818,"failed":0} ------------------------------ -• [0.091 seconds] -[sig-auth] ServiceAccounts -test/e2e/auth/framework.go:23 - should run through the lifecycle of a ServiceAccount [Conformance] - test/e2e/auth/service_accounts.go:646 +• [SLOW TEST] [10.174 seconds] +[sig-apps] ReplicationController +test/e2e/apps/framework.go:23 + should serve a basic image on each replica with a public image [Conformance] + test/e2e/apps/rc.go:66 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-auth] ServiceAccounts + [BeforeEach] [sig-apps] ReplicationController test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:07:18.178 - Dec 14 16:07:18.178: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename svcaccounts 12/14/22 16:07:18.18 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:18.203 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:18.207 - [It] should run through the lifecycle of a ServiceAccount [Conformance] - test/e2e/auth/service_accounts.go:646 - STEP: creating a ServiceAccount 12/14/22 16:07:18.21 - STEP: watching for the ServiceAccount to be added 12/14/22 16:07:18.219 - STEP: patching the ServiceAccount 12/14/22 16:07:18.221 - STEP: finding ServiceAccount in list of all ServiceAccounts (by LabelSelector) 12/14/22 16:07:18.231 - STEP: deleting the ServiceAccount 12/14/22 16:07:18.235 - [AfterEach] [sig-auth] ServiceAccounts + STEP: Creating a kubernetes client 02/25/23 10:11:28.629 + Feb 25 10:11:28.629: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename replication-controller 02/25/23 10:11:28.631 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:28.67 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:28.68 + [BeforeEach] [sig-apps] ReplicationController + test/e2e/apps/rc.go:56 + [It] should serve a basic image on each replica with a public image [Conformance] + test/e2e/apps/rc.go:66 + STEP: Creating replication controller my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f 02/25/23 10:11:28.686 + Feb 25 10:11:28.705: INFO: Pod name my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f: Found 0 pods out of 1 + Feb 25 10:11:33.717: INFO: Pod name my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f: Found 1 pods out of 1 + Feb 25 10:11:33.717: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f" are running + Feb 25 10:11:33.717: INFO: Waiting up to 5m0s for pod "my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f-ld2bv" in namespace "replication-controller-5303" to be "running" + Feb 25 10:11:33.728: INFO: Pod "my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f-ld2bv": Phase="Running", Reason="", readiness=true. Elapsed: 10.776153ms + Feb 25 10:11:33.728: INFO: Pod "my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f-ld2bv" satisfied condition "running" + Feb 25 10:11:33.728: INFO: Pod "my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f-ld2bv" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:11:28 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:11:30 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:11:30 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2023-02-25 10:11:28 +0000 UTC Reason: Message:}]) + Feb 25 10:11:33.728: INFO: Trying to dial the pod + Feb 25 10:11:38.763: INFO: Controller my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f: Got expected result from replica 1 [my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f-ld2bv]: "my-hostname-basic-176a2ab7-b05f-457d-ab25-652b66426d6f-ld2bv", 1 of 1 required successes so far + [AfterEach] [sig-apps] ReplicationController test/e2e/framework/framework.go:187 - Dec 14 16:07:18.254: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "svcaccounts-9374" for this suite. 12/14/22 16:07:18.259 + Feb 25 10:11:38.764: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "replication-controller-5303" for this suite. 02/25/23 10:11:38.775 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSS +SSSSSSSS ------------------------------ -[sig-node] ConfigMap - should be consumable via environment variable [NodeConformance] [Conformance] - test/e2e/common/node/configmap.go:44 -[BeforeEach] [sig-node] ConfigMap +[sig-api-machinery] Namespaces [Serial] + should apply changes to a namespace status [Conformance] + test/e2e/apimachinery/namespace.go:298 +[BeforeEach] [sig-api-machinery] Namespaces [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:07:18.274 -Dec 14 16:07:18.275: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename configmap 12/14/22 16:07:18.285 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:18.312 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:18.316 -[It] should be consumable via environment variable [NodeConformance] [Conformance] - test/e2e/common/node/configmap.go:44 -STEP: Creating configMap configmap-2536/configmap-test-b61bf4f3-6a7d-41a3-b039-213088cda3ef 12/14/22 16:07:18.32 -STEP: Creating a pod to test consume configMaps 12/14/22 16:07:18.326 -Dec 14 16:07:18.340: INFO: Waiting up to 5m0s for pod "pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef" in namespace "configmap-2536" to be "Succeeded or Failed" -Dec 14 16:07:18.344: INFO: Pod "pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef": Phase="Pending", Reason="", readiness=false. Elapsed: 3.812423ms -Dec 14 16:07:20.360: INFO: Pod "pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019411337s -Dec 14 16:07:22.353: INFO: Pod "pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012600287s -STEP: Saw pod success 12/14/22 16:07:22.353 -Dec 14 16:07:22.354: INFO: Pod "pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef" satisfied condition "Succeeded or Failed" -Dec 14 16:07:22.360: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef container env-test: -STEP: delete the pod 12/14/22 16:07:22.373 -Dec 14 16:07:22.393: INFO: Waiting for pod pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef to disappear -Dec 14 16:07:22.397: INFO: Pod pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef no longer exists -[AfterEach] [sig-node] ConfigMap +STEP: Creating a kubernetes client 02/25/23 10:11:38.804 +Feb 25 10:11:38.806: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename namespaces 02/25/23 10:11:38.814 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:38.853 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:38.86 +[It] should apply changes to a namespace status [Conformance] + test/e2e/apimachinery/namespace.go:298 +STEP: Read namespace status 02/25/23 10:11:38.886 +Feb 25 10:11:38.903: INFO: Status: v1.NamespaceStatus{Phase:"Active", Conditions:[]v1.NamespaceCondition(nil)} +STEP: Patch namespace status 02/25/23 10:11:38.903 +Feb 25 10:11:38.917: INFO: Status.Condition: v1.NamespaceCondition{Type:"StatusPatch", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Patched by an e2e test"} +STEP: Update namespace status 02/25/23 10:11:38.917 +Feb 25 10:11:38.944: INFO: Status.Condition: v1.NamespaceCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Updated by an e2e test"} +[AfterEach] [sig-api-machinery] Namespaces [Serial] test/e2e/framework/framework.go:187 -Dec 14 16:07:22.398: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-2536" for this suite. 12/14/22 16:07:22.404 -{"msg":"PASSED [sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance]","completed":87,"skipped":1542,"failed":0} +Feb 25 10:11:38.944: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "namespaces-856" for this suite. 02/25/23 10:11:38.954 +{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should apply changes to a namespace status [Conformance]","completed":89,"skipped":1826,"failed":0} ------------------------------ -• [4.140 seconds] -[sig-node] ConfigMap -test/e2e/common/node/framework.go:23 - should be consumable via environment variable [NodeConformance] [Conformance] - test/e2e/common/node/configmap.go:44 +• [0.166 seconds] +[sig-api-machinery] Namespaces [Serial] +test/e2e/apimachinery/framework.go:23 + should apply changes to a namespace status [Conformance] + test/e2e/apimachinery/namespace.go:298 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] ConfigMap + [BeforeEach] [sig-api-machinery] Namespaces [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:07:18.274 - Dec 14 16:07:18.275: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename configmap 12/14/22 16:07:18.285 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:18.312 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:18.316 - [It] should be consumable via environment variable [NodeConformance] [Conformance] - test/e2e/common/node/configmap.go:44 - STEP: Creating configMap configmap-2536/configmap-test-b61bf4f3-6a7d-41a3-b039-213088cda3ef 12/14/22 16:07:18.32 - STEP: Creating a pod to test consume configMaps 12/14/22 16:07:18.326 - Dec 14 16:07:18.340: INFO: Waiting up to 5m0s for pod "pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef" in namespace "configmap-2536" to be "Succeeded or Failed" - Dec 14 16:07:18.344: INFO: Pod "pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef": Phase="Pending", Reason="", readiness=false. Elapsed: 3.812423ms - Dec 14 16:07:20.360: INFO: Pod "pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019411337s - Dec 14 16:07:22.353: INFO: Pod "pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012600287s - STEP: Saw pod success 12/14/22 16:07:22.353 - Dec 14 16:07:22.354: INFO: Pod "pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef" satisfied condition "Succeeded or Failed" - Dec 14 16:07:22.360: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef container env-test: - STEP: delete the pod 12/14/22 16:07:22.373 - Dec 14 16:07:22.393: INFO: Waiting for pod pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef to disappear - Dec 14 16:07:22.397: INFO: Pod pod-configmaps-f92695c8-ec6c-45e2-a941-f927ec7aaaef no longer exists - [AfterEach] [sig-node] ConfigMap + STEP: Creating a kubernetes client 02/25/23 10:11:38.804 + Feb 25 10:11:38.806: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename namespaces 02/25/23 10:11:38.814 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:38.853 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:38.86 + [It] should apply changes to a namespace status [Conformance] + test/e2e/apimachinery/namespace.go:298 + STEP: Read namespace status 02/25/23 10:11:38.886 + Feb 25 10:11:38.903: INFO: Status: v1.NamespaceStatus{Phase:"Active", Conditions:[]v1.NamespaceCondition(nil)} + STEP: Patch namespace status 02/25/23 10:11:38.903 + Feb 25 10:11:38.917: INFO: Status.Condition: v1.NamespaceCondition{Type:"StatusPatch", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Patched by an e2e test"} + STEP: Update namespace status 02/25/23 10:11:38.917 + Feb 25 10:11:38.944: INFO: Status.Condition: v1.NamespaceCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Updated by an e2e test"} + [AfterEach] [sig-api-machinery] Namespaces [Serial] test/e2e/framework/framework.go:187 - Dec 14 16:07:22.398: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "configmap-2536" for this suite. 12/14/22 16:07:22.404 + Feb 25 10:11:38.944: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "namespaces-856" for this suite. 02/25/23 10:11:38.954 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSS +SSSSSSSSSSSS ------------------------------ -[sig-auth] ServiceAccounts - should mount projected service account token [Conformance] - test/e2e/auth/service_accounts.go:272 -[BeforeEach] [sig-auth] ServiceAccounts +[sig-apps] ReplicaSet + should list and delete a collection of ReplicaSets [Conformance] + test/e2e/apps/replica_set.go:165 +[BeforeEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:07:22.421 -Dec 14 16:07:22.422: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename svcaccounts 12/14/22 16:07:22.424 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:22.45 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:22.454 -[It] should mount projected service account token [Conformance] - test/e2e/auth/service_accounts.go:272 -STEP: Creating a pod to test service account token: 12/14/22 16:07:22.459 -Dec 14 16:07:22.471: INFO: Waiting up to 5m0s for pod "test-pod-77066820-667b-411c-8532-f4bb9548373c" in namespace "svcaccounts-4719" to be "Succeeded or Failed" -Dec 14 16:07:22.477: INFO: Pod "test-pod-77066820-667b-411c-8532-f4bb9548373c": Phase="Pending", Reason="", readiness=false. Elapsed: 6.712341ms -Dec 14 16:07:24.492: INFO: Pod "test-pod-77066820-667b-411c-8532-f4bb9548373c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020948349s -Dec 14 16:07:26.489: INFO: Pod "test-pod-77066820-667b-411c-8532-f4bb9548373c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018477179s -STEP: Saw pod success 12/14/22 16:07:26.489 -Dec 14 16:07:26.490: INFO: Pod "test-pod-77066820-667b-411c-8532-f4bb9548373c" satisfied condition "Succeeded or Failed" -Dec 14 16:07:26.496: INFO: Trying to get logs from node kiedai4iakae-3 pod test-pod-77066820-667b-411c-8532-f4bb9548373c container agnhost-container: -STEP: delete the pod 12/14/22 16:07:26.507 -Dec 14 16:07:26.529: INFO: Waiting for pod test-pod-77066820-667b-411c-8532-f4bb9548373c to disappear -Dec 14 16:07:26.533: INFO: Pod test-pod-77066820-667b-411c-8532-f4bb9548373c no longer exists -[AfterEach] [sig-auth] ServiceAccounts +STEP: Creating a kubernetes client 02/25/23 10:11:38.973 +Feb 25 10:11:38.974: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename replicaset 02/25/23 10:11:38.977 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:39.021 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:39.028 +[It] should list and delete a collection of ReplicaSets [Conformance] + test/e2e/apps/replica_set.go:165 +STEP: Create a ReplicaSet 02/25/23 10:11:39.036 +STEP: Verify that the required pods have come up 02/25/23 10:11:39.058 +Feb 25 10:11:39.069: INFO: Pod name sample-pod: Found 0 pods out of 3 +Feb 25 10:11:44.100: INFO: Pod name sample-pod: Found 3 pods out of 3 +STEP: ensuring each pod is running 02/25/23 10:11:44.101 +Feb 25 10:11:44.111: INFO: Replica Status: {Replicas:3 FullyLabeledReplicas:3 ReadyReplicas:3 AvailableReplicas:3 ObservedGeneration:1 Conditions:[]} +STEP: Listing all ReplicaSets 02/25/23 10:11:44.112 +STEP: DeleteCollection of the ReplicaSets 02/25/23 10:11:44.132 +STEP: After DeleteCollection verify that ReplicaSets have been deleted 02/25/23 10:11:44.157 +[AfterEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:187 -Dec 14 16:07:26.533: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "svcaccounts-4719" for this suite. 12/14/22 16:07:26.542 -{"msg":"PASSED [sig-auth] ServiceAccounts should mount projected service account token [Conformance]","completed":88,"skipped":1555,"failed":0} +Feb 25 10:11:44.166: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "replicaset-7248" for this suite. 02/25/23 10:11:44.18 +{"msg":"PASSED [sig-apps] ReplicaSet should list and delete a collection of ReplicaSets [Conformance]","completed":90,"skipped":1838,"failed":0} ------------------------------ -• [4.133 seconds] -[sig-auth] ServiceAccounts -test/e2e/auth/framework.go:23 - should mount projected service account token [Conformance] - test/e2e/auth/service_accounts.go:272 +• [SLOW TEST] [5.226 seconds] +[sig-apps] ReplicaSet +test/e2e/apps/framework.go:23 + should list and delete a collection of ReplicaSets [Conformance] + test/e2e/apps/replica_set.go:165 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-auth] ServiceAccounts + [BeforeEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:07:22.421 - Dec 14 16:07:22.422: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename svcaccounts 12/14/22 16:07:22.424 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:22.45 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:22.454 - [It] should mount projected service account token [Conformance] - test/e2e/auth/service_accounts.go:272 - STEP: Creating a pod to test service account token: 12/14/22 16:07:22.459 - Dec 14 16:07:22.471: INFO: Waiting up to 5m0s for pod "test-pod-77066820-667b-411c-8532-f4bb9548373c" in namespace "svcaccounts-4719" to be "Succeeded or Failed" - Dec 14 16:07:22.477: INFO: Pod "test-pod-77066820-667b-411c-8532-f4bb9548373c": Phase="Pending", Reason="", readiness=false. Elapsed: 6.712341ms - Dec 14 16:07:24.492: INFO: Pod "test-pod-77066820-667b-411c-8532-f4bb9548373c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020948349s - Dec 14 16:07:26.489: INFO: Pod "test-pod-77066820-667b-411c-8532-f4bb9548373c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018477179s - STEP: Saw pod success 12/14/22 16:07:26.489 - Dec 14 16:07:26.490: INFO: Pod "test-pod-77066820-667b-411c-8532-f4bb9548373c" satisfied condition "Succeeded or Failed" - Dec 14 16:07:26.496: INFO: Trying to get logs from node kiedai4iakae-3 pod test-pod-77066820-667b-411c-8532-f4bb9548373c container agnhost-container: - STEP: delete the pod 12/14/22 16:07:26.507 - Dec 14 16:07:26.529: INFO: Waiting for pod test-pod-77066820-667b-411c-8532-f4bb9548373c to disappear - Dec 14 16:07:26.533: INFO: Pod test-pod-77066820-667b-411c-8532-f4bb9548373c no longer exists - [AfterEach] [sig-auth] ServiceAccounts + STEP: Creating a kubernetes client 02/25/23 10:11:38.973 + Feb 25 10:11:38.974: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename replicaset 02/25/23 10:11:38.977 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:39.021 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:39.028 + [It] should list and delete a collection of ReplicaSets [Conformance] + test/e2e/apps/replica_set.go:165 + STEP: Create a ReplicaSet 02/25/23 10:11:39.036 + STEP: Verify that the required pods have come up 02/25/23 10:11:39.058 + Feb 25 10:11:39.069: INFO: Pod name sample-pod: Found 0 pods out of 3 + Feb 25 10:11:44.100: INFO: Pod name sample-pod: Found 3 pods out of 3 + STEP: ensuring each pod is running 02/25/23 10:11:44.101 + Feb 25 10:11:44.111: INFO: Replica Status: {Replicas:3 FullyLabeledReplicas:3 ReadyReplicas:3 AvailableReplicas:3 ObservedGeneration:1 Conditions:[]} + STEP: Listing all ReplicaSets 02/25/23 10:11:44.112 + STEP: DeleteCollection of the ReplicaSets 02/25/23 10:11:44.132 + STEP: After DeleteCollection verify that ReplicaSets have been deleted 02/25/23 10:11:44.157 + [AfterEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:187 - Dec 14 16:07:26.533: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "svcaccounts-4719" for this suite. 12/14/22 16:07:26.542 + Feb 25 10:11:44.166: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "replicaset-7248" for this suite. 02/25/23 10:11:44.18 << End Captured GinkgoWriter Output ------------------------------ -SS ------------------------------- -[sig-node] Security Context - should support pod.Spec.SecurityContext.RunAsUser And pod.Spec.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] - test/e2e/node/security_context.go:97 -[BeforeEach] [sig-node] Security Context +[sig-apps] Deployment + RecreateDeployment should delete old pods and create new ones [Conformance] + test/e2e/apps/deployment.go:113 +[BeforeEach] [sig-apps] Deployment test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:07:26.557 -Dec 14 16:07:26.557: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename security-context 12/14/22 16:07:26.56 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:26.586 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:26.589 -[It] should support pod.Spec.SecurityContext.RunAsUser And pod.Spec.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] - test/e2e/node/security_context.go:97 -STEP: Creating a pod to test pod.Spec.SecurityContext.RunAsUser 12/14/22 16:07:26.593 -Dec 14 16:07:26.609: INFO: Waiting up to 5m0s for pod "security-context-f7fc9783-6c79-4e04-90c3-b6c810158349" in namespace "security-context-37" to be "Succeeded or Failed" -Dec 14 16:07:26.615: INFO: Pod "security-context-f7fc9783-6c79-4e04-90c3-b6c810158349": Phase="Pending", Reason="", readiness=false. Elapsed: 6.591455ms -Dec 14 16:07:28.627: INFO: Pod "security-context-f7fc9783-6c79-4e04-90c3-b6c810158349": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018675232s -Dec 14 16:07:30.625: INFO: Pod "security-context-f7fc9783-6c79-4e04-90c3-b6c810158349": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015967766s -STEP: Saw pod success 12/14/22 16:07:30.625 -Dec 14 16:07:30.625: INFO: Pod "security-context-f7fc9783-6c79-4e04-90c3-b6c810158349" satisfied condition "Succeeded or Failed" -Dec 14 16:07:30.630: INFO: Trying to get logs from node kiedai4iakae-3 pod security-context-f7fc9783-6c79-4e04-90c3-b6c810158349 container test-container: -STEP: delete the pod 12/14/22 16:07:30.64 -Dec 14 16:07:30.655: INFO: Waiting for pod security-context-f7fc9783-6c79-4e04-90c3-b6c810158349 to disappear -Dec 14 16:07:30.659: INFO: Pod security-context-f7fc9783-6c79-4e04-90c3-b6c810158349 no longer exists -[AfterEach] [sig-node] Security Context +STEP: Creating a kubernetes client 02/25/23 10:11:44.2 +Feb 25 10:11:44.201: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename deployment 02/25/23 10:11:44.21 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:44.408 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:44.417 +[BeforeEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:91 +[It] RecreateDeployment should delete old pods and create new ones [Conformance] + test/e2e/apps/deployment.go:113 +Feb 25 10:11:44.424: INFO: Creating deployment "test-recreate-deployment" +Feb 25 10:11:44.448: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1 +Feb 25 10:11:44.464: INFO: new replicaset for deployment "test-recreate-deployment" is yet to be created +Feb 25 10:11:46.484: INFO: Waiting deployment "test-recreate-deployment" to complete +Feb 25 10:11:46.504: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 11, 44, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 11, 44, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 11, 44, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 11, 44, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-7d8b6f647f\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:11:48.513: INFO: Triggering a new rollout for deployment "test-recreate-deployment" +Feb 25 10:11:48.535: INFO: Updating deployment test-recreate-deployment +Feb 25 10:11:48.535: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods +[AfterEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:84 +Feb 25 10:11:48.942: INFO: Deployment "test-recreate-deployment": +&Deployment{ObjectMeta:{test-recreate-deployment deployment-4757 87b10a24-1a5e-4805-9378-4824b74f3905 16052 2 2023-02-25 10:11:44 +0000 UTC map[name:sample-pod-3] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:replicas":{},"f:unavailableReplicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0050a9c68 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2023-02-25 10:11:48 +0000 UTC,LastTransitionTime:2023-02-25 10:11:48 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "test-recreate-deployment-9d58999df" is progressing.,LastUpdateTime:2023-02-25 10:11:48 +0000 UTC,LastTransitionTime:2023-02-25 10:11:44 +0000 UTC,},},ReadyReplicas:0,CollisionCount:nil,},} + +Feb 25 10:11:48.976: INFO: New ReplicaSet "test-recreate-deployment-9d58999df" of Deployment "test-recreate-deployment": +&ReplicaSet{ObjectMeta:{test-recreate-deployment-9d58999df deployment-4757 13545e92-402f-4ecc-aadf-ecc7b5e082ad 16051 1 2023-02-25 10:11:48 +0000 UTC map[name:sample-pod-3 pod-template-hash:9d58999df] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-recreate-deployment 87b10a24-1a5e-4805-9378-4824b74f3905 0xc003164240 0xc003164241}] [] [{kube-controller-manager Update apps/v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"87b10a24-1a5e-4805-9378-4824b74f3905\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:status":{"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 9d58999df,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:9d58999df] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0031643f8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} +Feb 25 10:11:48.976: INFO: All old ReplicaSets of Deployment "test-recreate-deployment": +Feb 25 10:11:48.976: INFO: &ReplicaSet{ObjectMeta:{test-recreate-deployment-7d8b6f647f deployment-4757 74260362-fc24-4101-b447-81cd398a835b 16038 2 2023-02-25 10:11:44 +0000 UTC map[name:sample-pod-3 pod-template-hash:7d8b6f647f] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-recreate-deployment 87b10a24-1a5e-4805-9378-4824b74f3905 0xc003164017 0xc003164018}] [] [{kube-controller-manager Update apps/v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"87b10a24-1a5e-4805-9378-4824b74f3905\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 7d8b6f647f,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:7d8b6f647f] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003164118 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} +Feb 25 10:11:48.991: INFO: Pod "test-recreate-deployment-9d58999df-c99rw" is not available: +&Pod{ObjectMeta:{test-recreate-deployment-9d58999df-c99rw test-recreate-deployment-9d58999df- deployment-4757 405f4e6d-b76d-4c4e-9cb5-c8b43252cd3f 16049 0 2023-02-25 10:11:48 +0000 UTC map[name:sample-pod-3 pod-template-hash:9d58999df] map[] [{apps/v1 ReplicaSet test-recreate-deployment-9d58999df 13545e92-402f-4ecc-aadf-ecc7b5e082ad 0xc003bf7f80 0xc003bf7f81}] [] [{kube-controller-manager Update v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"13545e92-402f-4ecc-aadf-ecc7b5e082ad\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-vdh2t,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vdh2t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:11:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:11:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:11:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:11:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 10:11:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +[AfterEach] [sig-apps] Deployment test/e2e/framework/framework.go:187 -Dec 14 16:07:30.659: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "security-context-37" for this suite. 12/14/22 16:07:30.666 -{"msg":"PASSED [sig-node] Security Context should support pod.Spec.SecurityContext.RunAsUser And pod.Spec.SecurityContext.RunAsGroup [LinuxOnly] [Conformance]","completed":89,"skipped":1557,"failed":0} +Feb 25 10:11:48.991: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "deployment-4757" for this suite. 02/25/23 10:11:49.012 +{"msg":"PASSED [sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance]","completed":91,"skipped":1838,"failed":0} ------------------------------ -• [4.117 seconds] -[sig-node] Security Context -test/e2e/node/framework.go:23 - should support pod.Spec.SecurityContext.RunAsUser And pod.Spec.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] - test/e2e/node/security_context.go:97 +• [4.855 seconds] +[sig-apps] Deployment +test/e2e/apps/framework.go:23 + RecreateDeployment should delete old pods and create new ones [Conformance] + test/e2e/apps/deployment.go:113 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Security Context + [BeforeEach] [sig-apps] Deployment test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:07:26.557 - Dec 14 16:07:26.557: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename security-context 12/14/22 16:07:26.56 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:26.586 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:26.589 - [It] should support pod.Spec.SecurityContext.RunAsUser And pod.Spec.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] - test/e2e/node/security_context.go:97 - STEP: Creating a pod to test pod.Spec.SecurityContext.RunAsUser 12/14/22 16:07:26.593 - Dec 14 16:07:26.609: INFO: Waiting up to 5m0s for pod "security-context-f7fc9783-6c79-4e04-90c3-b6c810158349" in namespace "security-context-37" to be "Succeeded or Failed" - Dec 14 16:07:26.615: INFO: Pod "security-context-f7fc9783-6c79-4e04-90c3-b6c810158349": Phase="Pending", Reason="", readiness=false. Elapsed: 6.591455ms - Dec 14 16:07:28.627: INFO: Pod "security-context-f7fc9783-6c79-4e04-90c3-b6c810158349": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018675232s - Dec 14 16:07:30.625: INFO: Pod "security-context-f7fc9783-6c79-4e04-90c3-b6c810158349": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015967766s - STEP: Saw pod success 12/14/22 16:07:30.625 - Dec 14 16:07:30.625: INFO: Pod "security-context-f7fc9783-6c79-4e04-90c3-b6c810158349" satisfied condition "Succeeded or Failed" - Dec 14 16:07:30.630: INFO: Trying to get logs from node kiedai4iakae-3 pod security-context-f7fc9783-6c79-4e04-90c3-b6c810158349 container test-container: - STEP: delete the pod 12/14/22 16:07:30.64 - Dec 14 16:07:30.655: INFO: Waiting for pod security-context-f7fc9783-6c79-4e04-90c3-b6c810158349 to disappear - Dec 14 16:07:30.659: INFO: Pod security-context-f7fc9783-6c79-4e04-90c3-b6c810158349 no longer exists - [AfterEach] [sig-node] Security Context + STEP: Creating a kubernetes client 02/25/23 10:11:44.2 + Feb 25 10:11:44.201: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename deployment 02/25/23 10:11:44.21 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:44.408 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:44.417 + [BeforeEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:91 + [It] RecreateDeployment should delete old pods and create new ones [Conformance] + test/e2e/apps/deployment.go:113 + Feb 25 10:11:44.424: INFO: Creating deployment "test-recreate-deployment" + Feb 25 10:11:44.448: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1 + Feb 25 10:11:44.464: INFO: new replicaset for deployment "test-recreate-deployment" is yet to be created + Feb 25 10:11:46.484: INFO: Waiting deployment "test-recreate-deployment" to complete + Feb 25 10:11:46.504: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 11, 44, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 11, 44, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 11, 44, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 11, 44, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-recreate-deployment-7d8b6f647f\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:11:48.513: INFO: Triggering a new rollout for deployment "test-recreate-deployment" + Feb 25 10:11:48.535: INFO: Updating deployment test-recreate-deployment + Feb 25 10:11:48.535: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods + [AfterEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:84 + Feb 25 10:11:48.942: INFO: Deployment "test-recreate-deployment": + &Deployment{ObjectMeta:{test-recreate-deployment deployment-4757 87b10a24-1a5e-4805-9378-4824b74f3905 16052 2 2023-02-25 10:11:44 +0000 UTC map[name:sample-pod-3] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:replicas":{},"f:unavailableReplicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0050a9c68 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2023-02-25 10:11:48 +0000 UTC,LastTransitionTime:2023-02-25 10:11:48 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "test-recreate-deployment-9d58999df" is progressing.,LastUpdateTime:2023-02-25 10:11:48 +0000 UTC,LastTransitionTime:2023-02-25 10:11:44 +0000 UTC,},},ReadyReplicas:0,CollisionCount:nil,},} + + Feb 25 10:11:48.976: INFO: New ReplicaSet "test-recreate-deployment-9d58999df" of Deployment "test-recreate-deployment": + &ReplicaSet{ObjectMeta:{test-recreate-deployment-9d58999df deployment-4757 13545e92-402f-4ecc-aadf-ecc7b5e082ad 16051 1 2023-02-25 10:11:48 +0000 UTC map[name:sample-pod-3 pod-template-hash:9d58999df] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-recreate-deployment 87b10a24-1a5e-4805-9378-4824b74f3905 0xc003164240 0xc003164241}] [] [{kube-controller-manager Update apps/v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"87b10a24-1a5e-4805-9378-4824b74f3905\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:status":{"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 9d58999df,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:9d58999df] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0031643f8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} + Feb 25 10:11:48.976: INFO: All old ReplicaSets of Deployment "test-recreate-deployment": + Feb 25 10:11:48.976: INFO: &ReplicaSet{ObjectMeta:{test-recreate-deployment-7d8b6f647f deployment-4757 74260362-fc24-4101-b447-81cd398a835b 16038 2 2023-02-25 10:11:44 +0000 UTC map[name:sample-pod-3 pod-template-hash:7d8b6f647f] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-recreate-deployment 87b10a24-1a5e-4805-9378-4824b74f3905 0xc003164017 0xc003164018}] [] [{kube-controller-manager Update apps/v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"87b10a24-1a5e-4805-9378-4824b74f3905\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 7d8b6f647f,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:7d8b6f647f] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003164118 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} + Feb 25 10:11:48.991: INFO: Pod "test-recreate-deployment-9d58999df-c99rw" is not available: + &Pod{ObjectMeta:{test-recreate-deployment-9d58999df-c99rw test-recreate-deployment-9d58999df- deployment-4757 405f4e6d-b76d-4c4e-9cb5-c8b43252cd3f 16049 0 2023-02-25 10:11:48 +0000 UTC map[name:sample-pod-3 pod-template-hash:9d58999df] map[] [{apps/v1 ReplicaSet test-recreate-deployment-9d58999df 13545e92-402f-4ecc-aadf-ecc7b5e082ad 0xc003bf7f80 0xc003bf7f81}] [] [{kube-controller-manager Update v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"13545e92-402f-4ecc-aadf-ecc7b5e082ad\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 10:11:48 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-vdh2t,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vdh2t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:11:48 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:11:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:11:48 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:11:48 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:,StartTime:2023-02-25 10:11:48 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + [AfterEach] [sig-apps] Deployment test/e2e/framework/framework.go:187 - Dec 14 16:07:30.659: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "security-context-37" for this suite. 12/14/22 16:07:30.666 + Feb 25 10:11:48.991: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "deployment-4757" for this suite. 02/25/23 10:11:49.012 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +S ------------------------------ -[sig-storage] EmptyDir volumes - should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:206 -[BeforeEach] [sig-storage] EmptyDir volumes +[sig-node] PodTemplates + should delete a collection of pod templates [Conformance] + test/e2e/common/node/podtemplates.go:122 +[BeforeEach] [sig-node] PodTemplates test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:07:30.679 -Dec 14 16:07:30.679: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir 12/14/22 16:07:30.682 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:30.715 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:30.721 -[It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:206 -STEP: Creating a pod to test emptydir 0666 on node default medium 12/14/22 16:07:30.726 -Dec 14 16:07:30.737: INFO: Waiting up to 5m0s for pod "pod-b94030e7-b5f6-4d81-9304-eff0609bfa01" in namespace "emptydir-2544" to be "Succeeded or Failed" -Dec 14 16:07:30.741: INFO: Pod "pod-b94030e7-b5f6-4d81-9304-eff0609bfa01": Phase="Pending", Reason="", readiness=false. Elapsed: 3.327413ms -Dec 14 16:07:32.749: INFO: Pod "pod-b94030e7-b5f6-4d81-9304-eff0609bfa01": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011663424s -Dec 14 16:07:34.750: INFO: Pod "pod-b94030e7-b5f6-4d81-9304-eff0609bfa01": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012890753s -STEP: Saw pod success 12/14/22 16:07:34.75 -Dec 14 16:07:34.751: INFO: Pod "pod-b94030e7-b5f6-4d81-9304-eff0609bfa01" satisfied condition "Succeeded or Failed" -Dec 14 16:07:34.758: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-b94030e7-b5f6-4d81-9304-eff0609bfa01 container test-container: -STEP: delete the pod 12/14/22 16:07:34.771 -Dec 14 16:07:34.791: INFO: Waiting for pod pod-b94030e7-b5f6-4d81-9304-eff0609bfa01 to disappear -Dec 14 16:07:34.795: INFO: Pod pod-b94030e7-b5f6-4d81-9304-eff0609bfa01 no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +STEP: Creating a kubernetes client 02/25/23 10:11:49.062 +Feb 25 10:11:49.062: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename podtemplate 02/25/23 10:11:49.065 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:49.468 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:49.485 +[It] should delete a collection of pod templates [Conformance] + test/e2e/common/node/podtemplates.go:122 +STEP: Create set of pod templates 02/25/23 10:11:49.522 +Feb 25 10:11:49.577: INFO: created test-podtemplate-1 +Feb 25 10:11:49.592: INFO: created test-podtemplate-2 +Feb 25 10:11:49.651: INFO: created test-podtemplate-3 +STEP: get a list of pod templates with a label in the current namespace 02/25/23 10:11:49.663 +STEP: delete collection of pod templates 02/25/23 10:11:49.723 +Feb 25 10:11:49.724: INFO: requesting DeleteCollection of pod templates +STEP: check that the list of pod templates matches the requested quantity 02/25/23 10:11:49.771 +Feb 25 10:11:49.771: INFO: requesting list of pod templates to confirm quantity +[AfterEach] [sig-node] PodTemplates test/e2e/framework/framework.go:187 -Dec 14 16:07:34.795: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-2544" for this suite. 12/14/22 16:07:34.802 -{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","completed":90,"skipped":1597,"failed":0} +Feb 25 10:11:49.779: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "podtemplate-3937" for this suite. 02/25/23 10:11:49.79 +{"msg":"PASSED [sig-node] PodTemplates should delete a collection of pod templates [Conformance]","completed":92,"skipped":1839,"failed":0} ------------------------------ -• [4.133 seconds] -[sig-storage] EmptyDir volumes -test/e2e/common/storage/framework.go:23 - should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:206 +• [0.746 seconds] +[sig-node] PodTemplates +test/e2e/common/node/framework.go:23 + should delete a collection of pod templates [Conformance] + test/e2e/common/node/podtemplates.go:122 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir volumes + [BeforeEach] [sig-node] PodTemplates test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:07:30.679 - Dec 14 16:07:30.679: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir 12/14/22 16:07:30.682 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:30.715 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:30.721 - [It] should support (non-root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:206 - STEP: Creating a pod to test emptydir 0666 on node default medium 12/14/22 16:07:30.726 - Dec 14 16:07:30.737: INFO: Waiting up to 5m0s for pod "pod-b94030e7-b5f6-4d81-9304-eff0609bfa01" in namespace "emptydir-2544" to be "Succeeded or Failed" - Dec 14 16:07:30.741: INFO: Pod "pod-b94030e7-b5f6-4d81-9304-eff0609bfa01": Phase="Pending", Reason="", readiness=false. Elapsed: 3.327413ms - Dec 14 16:07:32.749: INFO: Pod "pod-b94030e7-b5f6-4d81-9304-eff0609bfa01": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011663424s - Dec 14 16:07:34.750: INFO: Pod "pod-b94030e7-b5f6-4d81-9304-eff0609bfa01": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012890753s - STEP: Saw pod success 12/14/22 16:07:34.75 - Dec 14 16:07:34.751: INFO: Pod "pod-b94030e7-b5f6-4d81-9304-eff0609bfa01" satisfied condition "Succeeded or Failed" - Dec 14 16:07:34.758: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-b94030e7-b5f6-4d81-9304-eff0609bfa01 container test-container: - STEP: delete the pod 12/14/22 16:07:34.771 - Dec 14 16:07:34.791: INFO: Waiting for pod pod-b94030e7-b5f6-4d81-9304-eff0609bfa01 to disappear - Dec 14 16:07:34.795: INFO: Pod pod-b94030e7-b5f6-4d81-9304-eff0609bfa01 no longer exists - [AfterEach] [sig-storage] EmptyDir volumes + STEP: Creating a kubernetes client 02/25/23 10:11:49.062 + Feb 25 10:11:49.062: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename podtemplate 02/25/23 10:11:49.065 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:49.468 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:49.485 + [It] should delete a collection of pod templates [Conformance] + test/e2e/common/node/podtemplates.go:122 + STEP: Create set of pod templates 02/25/23 10:11:49.522 + Feb 25 10:11:49.577: INFO: created test-podtemplate-1 + Feb 25 10:11:49.592: INFO: created test-podtemplate-2 + Feb 25 10:11:49.651: INFO: created test-podtemplate-3 + STEP: get a list of pod templates with a label in the current namespace 02/25/23 10:11:49.663 + STEP: delete collection of pod templates 02/25/23 10:11:49.723 + Feb 25 10:11:49.724: INFO: requesting DeleteCollection of pod templates + STEP: check that the list of pod templates matches the requested quantity 02/25/23 10:11:49.771 + Feb 25 10:11:49.771: INFO: requesting list of pod templates to confirm quantity + [AfterEach] [sig-node] PodTemplates test/e2e/framework/framework.go:187 - Dec 14 16:07:34.795: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-2544" for this suite. 12/14/22 16:07:34.802 + Feb 25 10:11:49.779: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "podtemplate-3937" for this suite. 02/25/23 10:11:49.79 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSS +SSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:116 -[BeforeEach] [sig-storage] EmptyDir volumes +[sig-api-machinery] Watchers + should observe an object deletion if it stops meeting the requirements of the selector [Conformance] + test/e2e/apimachinery/watch.go:257 +[BeforeEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:07:34.814 -Dec 14 16:07:34.814: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir 12/14/22 16:07:34.816 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:34.842 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:34.848 -[It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:116 -STEP: Creating a pod to test emptydir 0777 on tmpfs 12/14/22 16:07:34.852 -Dec 14 16:07:34.867: INFO: Waiting up to 5m0s for pod "pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae" in namespace "emptydir-3844" to be "Succeeded or Failed" -Dec 14 16:07:34.873: INFO: Pod "pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae": Phase="Pending", Reason="", readiness=false. Elapsed: 5.286073ms -Dec 14 16:07:36.882: INFO: Pod "pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015047494s -Dec 14 16:07:38.885: INFO: Pod "pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.017686692s -STEP: Saw pod success 12/14/22 16:07:38.885 -Dec 14 16:07:38.886: INFO: Pod "pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae" satisfied condition "Succeeded or Failed" -Dec 14 16:07:38.892: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae container test-container: -STEP: delete the pod 12/14/22 16:07:38.912 -Dec 14 16:07:38.934: INFO: Waiting for pod pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae to disappear -Dec 14 16:07:38.939: INFO: Pod pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +STEP: Creating a kubernetes client 02/25/23 10:11:49.814 +Feb 25 10:11:49.815: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename watch 02/25/23 10:11:49.82 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:49.879 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:49.886 +[It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance] + test/e2e/apimachinery/watch.go:257 +STEP: creating a watch on configmaps with a certain label 02/25/23 10:11:49.895 +STEP: creating a new configmap 02/25/23 10:11:49.898 +STEP: modifying the configmap once 02/25/23 10:11:49.923 +STEP: changing the label value of the configmap 02/25/23 10:11:49.957 +STEP: Expecting to observe a delete notification for the watched object 02/25/23 10:11:49.98 +Feb 25 10:11:49.980: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-1296 ba0e11b6-42e8-47a4-b48e-ff7242143549 16071 0 2023-02-25 10:11:49 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2023-02-25 10:11:49 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} +Feb 25 10:11:49.981: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-1296 ba0e11b6-42e8-47a4-b48e-ff7242143549 16073 0 2023-02-25 10:11:49 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2023-02-25 10:11:49 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} +Feb 25 10:11:49.982: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-1296 ba0e11b6-42e8-47a4-b48e-ff7242143549 16075 0 2023-02-25 10:11:49 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2023-02-25 10:11:49 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} +STEP: modifying the configmap a second time 02/25/23 10:11:49.982 +STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements 02/25/23 10:11:50.006 +STEP: changing the label value of the configmap back 02/25/23 10:12:00.007 +STEP: modifying the configmap a third time 02/25/23 10:12:00.032 +STEP: deleting the configmap 02/25/23 10:12:00.052 +STEP: Expecting to observe an add notification for the watched object when the label value was restored 02/25/23 10:12:00.067 +Feb 25 10:12:00.067: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-1296 ba0e11b6-42e8-47a4-b48e-ff7242143549 16141 0 2023-02-25 10:11:49 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2023-02-25 10:12:00 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} +Feb 25 10:12:00.068: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-1296 ba0e11b6-42e8-47a4-b48e-ff7242143549 16142 0 2023-02-25 10:11:49 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2023-02-25 10:12:00 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},Immutable:nil,} +Feb 25 10:12:00.068: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-1296 ba0e11b6-42e8-47a4-b48e-ff7242143549 16143 0 2023-02-25 10:11:49 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2023-02-25 10:12:00 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},Immutable:nil,} +[AfterEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:187 -Dec 14 16:07:38.940: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-3844" for this suite. 12/14/22 16:07:38.948 -{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","completed":91,"skipped":1618,"failed":0} +Feb 25 10:12:00.069: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "watch-1296" for this suite. 02/25/23 10:12:00.078 +{"msg":"PASSED [sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance]","completed":93,"skipped":1846,"failed":0} ------------------------------ -• [4.144 seconds] -[sig-storage] EmptyDir volumes -test/e2e/common/storage/framework.go:23 - should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:116 +• [SLOW TEST] [10.279 seconds] +[sig-api-machinery] Watchers +test/e2e/apimachinery/framework.go:23 + should observe an object deletion if it stops meeting the requirements of the selector [Conformance] + test/e2e/apimachinery/watch.go:257 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir volumes + [BeforeEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:07:34.814 - Dec 14 16:07:34.814: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir 12/14/22 16:07:34.816 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:34.842 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:34.848 - [It] should support (root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:116 - STEP: Creating a pod to test emptydir 0777 on tmpfs 12/14/22 16:07:34.852 - Dec 14 16:07:34.867: INFO: Waiting up to 5m0s for pod "pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae" in namespace "emptydir-3844" to be "Succeeded or Failed" - Dec 14 16:07:34.873: INFO: Pod "pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae": Phase="Pending", Reason="", readiness=false. Elapsed: 5.286073ms - Dec 14 16:07:36.882: INFO: Pod "pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015047494s - Dec 14 16:07:38.885: INFO: Pod "pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.017686692s - STEP: Saw pod success 12/14/22 16:07:38.885 - Dec 14 16:07:38.886: INFO: Pod "pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae" satisfied condition "Succeeded or Failed" - Dec 14 16:07:38.892: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae container test-container: - STEP: delete the pod 12/14/22 16:07:38.912 - Dec 14 16:07:38.934: INFO: Waiting for pod pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae to disappear - Dec 14 16:07:38.939: INFO: Pod pod-31fe40a8-d8ca-4de3-b7f7-a4d039cbddae no longer exists - [AfterEach] [sig-storage] EmptyDir volumes + STEP: Creating a kubernetes client 02/25/23 10:11:49.814 + Feb 25 10:11:49.815: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename watch 02/25/23 10:11:49.82 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:11:49.879 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:11:49.886 + [It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance] + test/e2e/apimachinery/watch.go:257 + STEP: creating a watch on configmaps with a certain label 02/25/23 10:11:49.895 + STEP: creating a new configmap 02/25/23 10:11:49.898 + STEP: modifying the configmap once 02/25/23 10:11:49.923 + STEP: changing the label value of the configmap 02/25/23 10:11:49.957 + STEP: Expecting to observe a delete notification for the watched object 02/25/23 10:11:49.98 + Feb 25 10:11:49.980: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-1296 ba0e11b6-42e8-47a4-b48e-ff7242143549 16071 0 2023-02-25 10:11:49 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2023-02-25 10:11:49 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} + Feb 25 10:11:49.981: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-1296 ba0e11b6-42e8-47a4-b48e-ff7242143549 16073 0 2023-02-25 10:11:49 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2023-02-25 10:11:49 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} + Feb 25 10:11:49.982: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-1296 ba0e11b6-42e8-47a4-b48e-ff7242143549 16075 0 2023-02-25 10:11:49 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2023-02-25 10:11:49 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} + STEP: modifying the configmap a second time 02/25/23 10:11:49.982 + STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements 02/25/23 10:11:50.006 + STEP: changing the label value of the configmap back 02/25/23 10:12:00.007 + STEP: modifying the configmap a third time 02/25/23 10:12:00.032 + STEP: deleting the configmap 02/25/23 10:12:00.052 + STEP: Expecting to observe an add notification for the watched object when the label value was restored 02/25/23 10:12:00.067 + Feb 25 10:12:00.067: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-1296 ba0e11b6-42e8-47a4-b48e-ff7242143549 16141 0 2023-02-25 10:11:49 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2023-02-25 10:12:00 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} + Feb 25 10:12:00.068: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-1296 ba0e11b6-42e8-47a4-b48e-ff7242143549 16142 0 2023-02-25 10:11:49 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2023-02-25 10:12:00 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},Immutable:nil,} + Feb 25 10:12:00.068: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-1296 ba0e11b6-42e8-47a4-b48e-ff7242143549 16143 0 2023-02-25 10:11:49 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2023-02-25 10:12:00 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},Immutable:nil,} + [AfterEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:187 - Dec 14 16:07:38.940: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-3844" for this suite. 12/14/22 16:07:38.948 + Feb 25 10:12:00.069: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "watch-1296" for this suite. 02/25/23 10:12:00.078 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSS +SS ------------------------------ -[sig-storage] EmptyDir volumes - should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:126 -[BeforeEach] [sig-storage] EmptyDir volumes +[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] + custom resource defaulting for requests and from storage works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:269 +[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:07:38.963 -Dec 14 16:07:38.963: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir 12/14/22 16:07:38.965 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:38.993 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:38.996 -[It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:126 -STEP: Creating a pod to test emptydir 0644 on tmpfs 12/14/22 16:07:39.003 -Dec 14 16:07:39.014: INFO: Waiting up to 5m0s for pod "pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e" in namespace "emptydir-4535" to be "Succeeded or Failed" -Dec 14 16:07:39.017: INFO: Pod "pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e": Phase="Pending", Reason="", readiness=false. Elapsed: 3.393463ms -Dec 14 16:07:41.028: INFO: Pod "pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014309695s -Dec 14 16:07:43.030: INFO: Pod "pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016059818s -STEP: Saw pod success 12/14/22 16:07:43.03 -Dec 14 16:07:43.031: INFO: Pod "pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e" satisfied condition "Succeeded or Failed" -Dec 14 16:07:43.036: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e container test-container: -STEP: delete the pod 12/14/22 16:07:43.048 -Dec 14 16:07:43.068: INFO: Waiting for pod pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e to disappear -Dec 14 16:07:43.074: INFO: Pod pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +STEP: Creating a kubernetes client 02/25/23 10:12:00.098 +Feb 25 10:12:00.099: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename custom-resource-definition 02/25/23 10:12:00.11 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:12:00.159 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:12:00.167 +[It] custom resource defaulting for requests and from storage works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:269 +Feb 25 10:12:00.173: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:07:43.075: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-4535" for this suite. 12/14/22 16:07:43.083 -{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","completed":92,"skipped":1639,"failed":0} +Feb 25 10:12:03.945: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "custom-resource-definition-4152" for this suite. 02/25/23 10:12:03.955 +{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] custom resource defaulting for requests and from storage works [Conformance]","completed":94,"skipped":1848,"failed":0} ------------------------------ -• [4.131 seconds] -[sig-storage] EmptyDir volumes -test/e2e/common/storage/framework.go:23 - should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:126 +• [3.883 seconds] +[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + custom resource defaulting for requests and from storage works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:269 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir volumes + [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:07:38.963 - Dec 14 16:07:38.963: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir 12/14/22 16:07:38.965 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:38.993 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:38.996 - [It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:126 - STEP: Creating a pod to test emptydir 0644 on tmpfs 12/14/22 16:07:39.003 - Dec 14 16:07:39.014: INFO: Waiting up to 5m0s for pod "pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e" in namespace "emptydir-4535" to be "Succeeded or Failed" - Dec 14 16:07:39.017: INFO: Pod "pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e": Phase="Pending", Reason="", readiness=false. Elapsed: 3.393463ms - Dec 14 16:07:41.028: INFO: Pod "pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014309695s - Dec 14 16:07:43.030: INFO: Pod "pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016059818s - STEP: Saw pod success 12/14/22 16:07:43.03 - Dec 14 16:07:43.031: INFO: Pod "pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e" satisfied condition "Succeeded or Failed" - Dec 14 16:07:43.036: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e container test-container: - STEP: delete the pod 12/14/22 16:07:43.048 - Dec 14 16:07:43.068: INFO: Waiting for pod pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e to disappear - Dec 14 16:07:43.074: INFO: Pod pod-26b3f3ad-b343-41b4-9cdd-d4be6b9c5b1e no longer exists - [AfterEach] [sig-storage] EmptyDir volumes + STEP: Creating a kubernetes client 02/25/23 10:12:00.098 + Feb 25 10:12:00.099: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename custom-resource-definition 02/25/23 10:12:00.11 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:12:00.159 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:12:00.167 + [It] custom resource defaulting for requests and from storage works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:269 + Feb 25 10:12:00.173: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:07:43.075: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-4535" for this suite. 12/14/22 16:07:43.083 + Feb 25 10:12:03.945: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "custom-resource-definition-4152" for this suite. 02/25/23 10:12:03.955 << End Captured GinkgoWriter Output ------------------------------ -[sig-apps] ReplicationController - should serve a basic image on each replica with a public image [Conformance] - test/e2e/apps/rc.go:66 -[BeforeEach] [sig-apps] ReplicationController +[sig-storage] Downward API volume + should update annotations on modification [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:161 +[BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:07:43.095 -Dec 14 16:07:43.095: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename replication-controller 12/14/22 16:07:43.102 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:43.137 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:43.14 -[BeforeEach] [sig-apps] ReplicationController - test/e2e/apps/rc.go:56 -[It] should serve a basic image on each replica with a public image [Conformance] - test/e2e/apps/rc.go:66 -STEP: Creating replication controller my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1 12/14/22 16:07:43.146 -Dec 14 16:07:43.159: INFO: Pod name my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1: Found 0 pods out of 1 -Dec 14 16:07:48.165: INFO: Pod name my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1: Found 1 pods out of 1 -Dec 14 16:07:48.166: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1" are running -Dec 14 16:07:48.166: INFO: Waiting up to 5m0s for pod "my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1-bgv6m" in namespace "replication-controller-7694" to be "running" -Dec 14 16:07:48.169: INFO: Pod "my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1-bgv6m": Phase="Running", Reason="", readiness=true. Elapsed: 3.679333ms -Dec 14 16:07:48.169: INFO: Pod "my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1-bgv6m" satisfied condition "running" -Dec 14 16:07:48.170: INFO: Pod "my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1-bgv6m" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 16:07:43 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 16:07:44 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 16:07:44 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 16:07:43 +0000 UTC Reason: Message:}]) -Dec 14 16:07:48.170: INFO: Trying to dial the pod -Dec 14 16:07:53.195: INFO: Controller my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1: Got expected result from replica 1 [my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1-bgv6m]: "my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1-bgv6m", 1 of 1 required successes so far -[AfterEach] [sig-apps] ReplicationController +STEP: Creating a kubernetes client 02/25/23 10:12:03.983 +Feb 25 10:12:03.984: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 10:12:03.989 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:12:04.042 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:12:04.051 +[BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 +[It] should update annotations on modification [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:161 +STEP: Creating the pod 02/25/23 10:12:04.061 +Feb 25 10:12:04.100: INFO: Waiting up to 5m0s for pod "annotationupdatecc6ebbad-78af-4e3c-9594-74800201325e" in namespace "downward-api-4022" to be "running and ready" +Feb 25 10:12:04.128: INFO: Pod "annotationupdatecc6ebbad-78af-4e3c-9594-74800201325e": Phase="Pending", Reason="", readiness=false. Elapsed: 28.222667ms +Feb 25 10:12:04.128: INFO: The phase of Pod annotationupdatecc6ebbad-78af-4e3c-9594-74800201325e is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:12:06.141: INFO: Pod "annotationupdatecc6ebbad-78af-4e3c-9594-74800201325e": Phase="Running", Reason="", readiness=true. Elapsed: 2.041602104s +Feb 25 10:12:06.142: INFO: The phase of Pod annotationupdatecc6ebbad-78af-4e3c-9594-74800201325e is Running (Ready = true) +Feb 25 10:12:06.142: INFO: Pod "annotationupdatecc6ebbad-78af-4e3c-9594-74800201325e" satisfied condition "running and ready" +Feb 25 10:12:06.688: INFO: Successfully updated pod "annotationupdatecc6ebbad-78af-4e3c-9594-74800201325e" +[AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 -Dec 14 16:07:53.196: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "replication-controller-7694" for this suite. 12/14/22 16:07:53.202 -{"msg":"PASSED [sig-apps] ReplicationController should serve a basic image on each replica with a public image [Conformance]","completed":93,"skipped":1639,"failed":0} +Feb 25 10:12:10.748: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-4022" for this suite. 02/25/23 10:12:10.759 +{"msg":"PASSED [sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance]","completed":95,"skipped":1848,"failed":0} ------------------------------ -• [SLOW TEST] [10.118 seconds] -[sig-apps] ReplicationController -test/e2e/apps/framework.go:23 - should serve a basic image on each replica with a public image [Conformance] - test/e2e/apps/rc.go:66 +• [SLOW TEST] [6.791 seconds] +[sig-storage] Downward API volume +test/e2e/common/storage/framework.go:23 + should update annotations on modification [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:161 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] ReplicationController + [BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:07:43.095 - Dec 14 16:07:43.095: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename replication-controller 12/14/22 16:07:43.102 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:43.137 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:43.14 - [BeforeEach] [sig-apps] ReplicationController - test/e2e/apps/rc.go:56 - [It] should serve a basic image on each replica with a public image [Conformance] - test/e2e/apps/rc.go:66 - STEP: Creating replication controller my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1 12/14/22 16:07:43.146 - Dec 14 16:07:43.159: INFO: Pod name my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1: Found 0 pods out of 1 - Dec 14 16:07:48.165: INFO: Pod name my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1: Found 1 pods out of 1 - Dec 14 16:07:48.166: INFO: Ensuring all pods for ReplicationController "my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1" are running - Dec 14 16:07:48.166: INFO: Waiting up to 5m0s for pod "my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1-bgv6m" in namespace "replication-controller-7694" to be "running" - Dec 14 16:07:48.169: INFO: Pod "my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1-bgv6m": Phase="Running", Reason="", readiness=true. Elapsed: 3.679333ms - Dec 14 16:07:48.169: INFO: Pod "my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1-bgv6m" satisfied condition "running" - Dec 14 16:07:48.170: INFO: Pod "my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1-bgv6m" is running (conditions: [{Type:Initialized Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 16:07:43 +0000 UTC Reason: Message:} {Type:Ready Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 16:07:44 +0000 UTC Reason: Message:} {Type:ContainersReady Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 16:07:44 +0000 UTC Reason: Message:} {Type:PodScheduled Status:True LastProbeTime:0001-01-01 00:00:00 +0000 UTC LastTransitionTime:2022-12-14 16:07:43 +0000 UTC Reason: Message:}]) - Dec 14 16:07:48.170: INFO: Trying to dial the pod - Dec 14 16:07:53.195: INFO: Controller my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1: Got expected result from replica 1 [my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1-bgv6m]: "my-hostname-basic-c2825fec-4500-4330-ba79-c855047a9fb1-bgv6m", 1 of 1 required successes so far - [AfterEach] [sig-apps] ReplicationController + STEP: Creating a kubernetes client 02/25/23 10:12:03.983 + Feb 25 10:12:03.984: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 10:12:03.989 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:12:04.042 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:12:04.051 + [BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 + [It] should update annotations on modification [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:161 + STEP: Creating the pod 02/25/23 10:12:04.061 + Feb 25 10:12:04.100: INFO: Waiting up to 5m0s for pod "annotationupdatecc6ebbad-78af-4e3c-9594-74800201325e" in namespace "downward-api-4022" to be "running and ready" + Feb 25 10:12:04.128: INFO: Pod "annotationupdatecc6ebbad-78af-4e3c-9594-74800201325e": Phase="Pending", Reason="", readiness=false. Elapsed: 28.222667ms + Feb 25 10:12:04.128: INFO: The phase of Pod annotationupdatecc6ebbad-78af-4e3c-9594-74800201325e is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:12:06.141: INFO: Pod "annotationupdatecc6ebbad-78af-4e3c-9594-74800201325e": Phase="Running", Reason="", readiness=true. Elapsed: 2.041602104s + Feb 25 10:12:06.142: INFO: The phase of Pod annotationupdatecc6ebbad-78af-4e3c-9594-74800201325e is Running (Ready = true) + Feb 25 10:12:06.142: INFO: Pod "annotationupdatecc6ebbad-78af-4e3c-9594-74800201325e" satisfied condition "running and ready" + Feb 25 10:12:06.688: INFO: Successfully updated pod "annotationupdatecc6ebbad-78af-4e3c-9594-74800201325e" + [AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 - Dec 14 16:07:53.196: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "replication-controller-7694" for this suite. 12/14/22 16:07:53.202 + Feb 25 10:12:10.748: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-4022" for this suite. 02/25/23 10:12:10.759 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +S ------------------------------ -[sig-node] NoExecuteTaintManager Single Pod [Serial] - removing taint cancels eviction [Disruptive] [Conformance] - test/e2e/node/taints.go:289 -[BeforeEach] [sig-node] NoExecuteTaintManager Single Pod [Serial] +[sig-node] Secrets + should patch a secret [Conformance] + test/e2e/common/node/secrets.go:153 +[BeforeEach] [sig-node] Secrets test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:07:53.233 -Dec 14 16:07:53.233: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename taint-single-pod 12/14/22 16:07:53.235 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:53.262 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:53.266 -[BeforeEach] [sig-node] NoExecuteTaintManager Single Pod [Serial] - test/e2e/node/taints.go:166 -Dec 14 16:07:53.270: INFO: Waiting up to 1m0s for all nodes to be ready -Dec 14 16:08:53.310: INFO: Waiting for terminating namespaces to be deleted... -[It] removing taint cancels eviction [Disruptive] [Conformance] - test/e2e/node/taints.go:289 -Dec 14 16:08:53.316: INFO: Starting informer... -STEP: Starting pod... 12/14/22 16:08:53.316 -Dec 14 16:08:53.544: INFO: Pod is running on kiedai4iakae-3. Tainting Node -STEP: Trying to apply a taint on the Node 12/14/22 16:08:53.544 -STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 12/14/22 16:08:53.578 -STEP: Waiting short time to make sure Pod is queued for deletion 12/14/22 16:08:53.586 -Dec 14 16:08:53.589: INFO: Pod wasn't evicted. Proceeding -Dec 14 16:08:53.589: INFO: Removing taint from Node -STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 12/14/22 16:08:53.663 -STEP: Waiting some time to make sure that toleration time passed. 12/14/22 16:08:53.681 -Dec 14 16:10:08.682: INFO: Pod wasn't evicted. Test successful -[AfterEach] [sig-node] NoExecuteTaintManager Single Pod [Serial] +STEP: Creating a kubernetes client 02/25/23 10:12:10.777 +Feb 25 10:12:10.778: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename secrets 02/25/23 10:12:10.783 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:12:10.823 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:12:10.829 +[It] should patch a secret [Conformance] + test/e2e/common/node/secrets.go:153 +STEP: creating a secret 02/25/23 10:12:10.834 +STEP: listing secrets in all namespaces to ensure that there are more than zero 02/25/23 10:12:10.845 +STEP: patching the secret 02/25/23 10:12:10.856 +STEP: deleting the secret using a LabelSelector 02/25/23 10:12:10.895 +STEP: listing secrets in all namespaces, searching for label name and value in patch 02/25/23 10:12:10.915 +[AfterEach] [sig-node] Secrets test/e2e/framework/framework.go:187 -Dec 14 16:10:08.688: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "taint-single-pod-9724" for this suite. 12/14/22 16:10:08.708 -{"msg":"PASSED [sig-node] NoExecuteTaintManager Single Pod [Serial] removing taint cancels eviction [Disruptive] [Conformance]","completed":94,"skipped":1692,"failed":0} +Feb 25 10:12:10.924: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-2984" for this suite. 02/25/23 10:12:10.933 +{"msg":"PASSED [sig-node] Secrets should patch a secret [Conformance]","completed":96,"skipped":1849,"failed":0} ------------------------------ -• [SLOW TEST] [135.487 seconds] -[sig-node] NoExecuteTaintManager Single Pod [Serial] -test/e2e/node/framework.go:23 - removing taint cancels eviction [Disruptive] [Conformance] - test/e2e/node/taints.go:289 +• [0.173 seconds] +[sig-node] Secrets +test/e2e/common/node/framework.go:23 + should patch a secret [Conformance] + test/e2e/common/node/secrets.go:153 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] NoExecuteTaintManager Single Pod [Serial] + [BeforeEach] [sig-node] Secrets test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:07:53.233 - Dec 14 16:07:53.233: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename taint-single-pod 12/14/22 16:07:53.235 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:07:53.262 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:07:53.266 - [BeforeEach] [sig-node] NoExecuteTaintManager Single Pod [Serial] - test/e2e/node/taints.go:166 - Dec 14 16:07:53.270: INFO: Waiting up to 1m0s for all nodes to be ready - Dec 14 16:08:53.310: INFO: Waiting for terminating namespaces to be deleted... - [It] removing taint cancels eviction [Disruptive] [Conformance] - test/e2e/node/taints.go:289 - Dec 14 16:08:53.316: INFO: Starting informer... - STEP: Starting pod... 12/14/22 16:08:53.316 - Dec 14 16:08:53.544: INFO: Pod is running on kiedai4iakae-3. Tainting Node - STEP: Trying to apply a taint on the Node 12/14/22 16:08:53.544 - STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 12/14/22 16:08:53.578 - STEP: Waiting short time to make sure Pod is queued for deletion 12/14/22 16:08:53.586 - Dec 14 16:08:53.589: INFO: Pod wasn't evicted. Proceeding - Dec 14 16:08:53.589: INFO: Removing taint from Node - STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 12/14/22 16:08:53.663 - STEP: Waiting some time to make sure that toleration time passed. 12/14/22 16:08:53.681 - Dec 14 16:10:08.682: INFO: Pod wasn't evicted. Test successful - [AfterEach] [sig-node] NoExecuteTaintManager Single Pod [Serial] + STEP: Creating a kubernetes client 02/25/23 10:12:10.777 + Feb 25 10:12:10.778: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename secrets 02/25/23 10:12:10.783 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:12:10.823 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:12:10.829 + [It] should patch a secret [Conformance] + test/e2e/common/node/secrets.go:153 + STEP: creating a secret 02/25/23 10:12:10.834 + STEP: listing secrets in all namespaces to ensure that there are more than zero 02/25/23 10:12:10.845 + STEP: patching the secret 02/25/23 10:12:10.856 + STEP: deleting the secret using a LabelSelector 02/25/23 10:12:10.895 + STEP: listing secrets in all namespaces, searching for label name and value in patch 02/25/23 10:12:10.915 + [AfterEach] [sig-node] Secrets test/e2e/framework/framework.go:187 - Dec 14 16:10:08.688: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "taint-single-pod-9724" for this suite. 12/14/22 16:10:08.708 + Feb 25 10:12:10.924: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "secrets-2984" for this suite. 02/25/23 10:12:10.933 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSS +SSSSSSSSSSSS ------------------------------ -[sig-apps] ReplicaSet - Replicaset should have a working scale subresource [Conformance] - test/e2e/apps/replica_set.go:143 -[BeforeEach] [sig-apps] ReplicaSet +[sig-node] Container Runtime blackbox test on terminated container + should report termination message if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:194 +[BeforeEach] [sig-node] Container Runtime test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:10:08.722 -Dec 14 16:10:08.723: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename replicaset 12/14/22 16:10:08.727 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:10:08.76 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:10:08.763 -[It] Replicaset should have a working scale subresource [Conformance] - test/e2e/apps/replica_set.go:143 -STEP: Creating replica set "test-rs" that asks for more than the allowed pod quota 12/14/22 16:10:08.766 -Dec 14 16:10:08.780: INFO: Pod name sample-pod: Found 0 pods out of 1 -Dec 14 16:10:13.792: INFO: Pod name sample-pod: Found 1 pods out of 1 -STEP: ensuring each pod is running 12/14/22 16:10:13.792 -STEP: getting scale subresource 12/14/22 16:10:13.792 -STEP: updating a scale subresource 12/14/22 16:10:13.798 -STEP: verifying the replicaset Spec.Replicas was modified 12/14/22 16:10:13.808 -STEP: Patch a scale subresource 12/14/22 16:10:13.813 -[AfterEach] [sig-apps] ReplicaSet +STEP: Creating a kubernetes client 02/25/23 10:12:10.952 +Feb 25 10:12:10.952: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-runtime 02/25/23 10:12:10.956 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:12:11.014 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:12:11.025 +[It] should report termination message if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:194 +STEP: create the container 02/25/23 10:12:11.035 +STEP: wait for the container to reach Succeeded 02/25/23 10:12:11.064 +STEP: get the container status 02/25/23 10:12:15.131 +STEP: the container should be terminated 02/25/23 10:12:15.136 +STEP: the termination message should be set 02/25/23 10:12:15.136 +Feb 25 10:12:15.137: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- +STEP: delete the container 02/25/23 10:12:15.137 +[AfterEach] [sig-node] Container Runtime test/e2e/framework/framework.go:187 -Dec 14 16:10:13.840: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "replicaset-2286" for this suite. 12/14/22 16:10:13.851 -{"msg":"PASSED [sig-apps] ReplicaSet Replicaset should have a working scale subresource [Conformance]","completed":95,"skipped":1702,"failed":0} +Feb 25 10:12:15.180: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-runtime-1093" for this suite. 02/25/23 10:12:15.192 +{"msg":"PASSED [sig-node] Container Runtime blackbox test on terminated container should report termination message if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]","completed":97,"skipped":1861,"failed":0} ------------------------------ -• [SLOW TEST] [5.144 seconds] -[sig-apps] ReplicaSet -test/e2e/apps/framework.go:23 - Replicaset should have a working scale subresource [Conformance] - test/e2e/apps/replica_set.go:143 +• [4.261 seconds] +[sig-node] Container Runtime +test/e2e/common/node/framework.go:23 + blackbox test + test/e2e/common/node/runtime.go:43 + on terminated container + test/e2e/common/node/runtime.go:136 + should report termination message if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:194 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] ReplicaSet + [BeforeEach] [sig-node] Container Runtime test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:10:08.722 - Dec 14 16:10:08.723: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename replicaset 12/14/22 16:10:08.727 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:10:08.76 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:10:08.763 - [It] Replicaset should have a working scale subresource [Conformance] - test/e2e/apps/replica_set.go:143 - STEP: Creating replica set "test-rs" that asks for more than the allowed pod quota 12/14/22 16:10:08.766 - Dec 14 16:10:08.780: INFO: Pod name sample-pod: Found 0 pods out of 1 - Dec 14 16:10:13.792: INFO: Pod name sample-pod: Found 1 pods out of 1 - STEP: ensuring each pod is running 12/14/22 16:10:13.792 - STEP: getting scale subresource 12/14/22 16:10:13.792 - STEP: updating a scale subresource 12/14/22 16:10:13.798 - STEP: verifying the replicaset Spec.Replicas was modified 12/14/22 16:10:13.808 - STEP: Patch a scale subresource 12/14/22 16:10:13.813 - [AfterEach] [sig-apps] ReplicaSet + STEP: Creating a kubernetes client 02/25/23 10:12:10.952 + Feb 25 10:12:10.952: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-runtime 02/25/23 10:12:10.956 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:12:11.014 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:12:11.025 + [It] should report termination message if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:194 + STEP: create the container 02/25/23 10:12:11.035 + STEP: wait for the container to reach Succeeded 02/25/23 10:12:11.064 + STEP: get the container status 02/25/23 10:12:15.131 + STEP: the container should be terminated 02/25/23 10:12:15.136 + STEP: the termination message should be set 02/25/23 10:12:15.136 + Feb 25 10:12:15.137: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- + STEP: delete the container 02/25/23 10:12:15.137 + [AfterEach] [sig-node] Container Runtime test/e2e/framework/framework.go:187 - Dec 14 16:10:13.840: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "replicaset-2286" for this suite. 12/14/22 16:10:13.851 + Feb 25 10:12:15.180: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-runtime-1093" for this suite. 02/25/23 10:12:15.192 << End Captured GinkgoWriter Output ------------------------------ -S +SSSSSSSSSS ------------------------------ -[sig-node] InitContainer [NodeConformance] - should invoke init containers on a RestartAlways pod [Conformance] - test/e2e/common/node/init_container.go:254 -[BeforeEach] [sig-node] InitContainer [NodeConformance] +[sig-architecture] Conformance Tests + should have at least two untainted nodes [Conformance] + test/e2e/architecture/conformance.go:38 +[BeforeEach] [sig-architecture] Conformance Tests test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:10:13.867 -Dec 14 16:10:13.867: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename init-container 12/14/22 16:10:13.871 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:10:13.914 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:10:13.917 -[BeforeEach] [sig-node] InitContainer [NodeConformance] - test/e2e/common/node/init_container.go:164 -[It] should invoke init containers on a RestartAlways pod [Conformance] - test/e2e/common/node/init_container.go:254 -STEP: creating the pod 12/14/22 16:10:13.919 -Dec 14 16:10:13.920: INFO: PodSpec: initContainers in spec.initContainers -[AfterEach] [sig-node] InitContainer [NodeConformance] +STEP: Creating a kubernetes client 02/25/23 10:12:15.218 +Feb 25 10:12:15.218: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename conformance-tests 02/25/23 10:12:15.221 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:12:15.26 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:12:15.265 +[It] should have at least two untainted nodes [Conformance] + test/e2e/architecture/conformance.go:38 +STEP: Getting node addresses 02/25/23 10:12:15.272 +Feb 25 10:12:15.273: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable +[AfterEach] [sig-architecture] Conformance Tests test/e2e/framework/framework.go:187 -Dec 14 16:10:17.146: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "init-container-6728" for this suite. 12/14/22 16:10:17.153 -{"msg":"PASSED [sig-node] InitContainer [NodeConformance] should invoke init containers on a RestartAlways pod [Conformance]","completed":96,"skipped":1703,"failed":0} +Feb 25 10:12:15.287: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "conformance-tests-6894" for this suite. 02/25/23 10:12:15.297 +{"msg":"PASSED [sig-architecture] Conformance Tests should have at least two untainted nodes [Conformance]","completed":98,"skipped":1871,"failed":0} ------------------------------ -• [3.295 seconds] -[sig-node] InitContainer [NodeConformance] -test/e2e/common/node/framework.go:23 - should invoke init containers on a RestartAlways pod [Conformance] - test/e2e/common/node/init_container.go:254 +• [0.097 seconds] +[sig-architecture] Conformance Tests +test/e2e/architecture/framework.go:23 + should have at least two untainted nodes [Conformance] + test/e2e/architecture/conformance.go:38 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] InitContainer [NodeConformance] + [BeforeEach] [sig-architecture] Conformance Tests test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:10:13.867 - Dec 14 16:10:13.867: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename init-container 12/14/22 16:10:13.871 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:10:13.914 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:10:13.917 - [BeforeEach] [sig-node] InitContainer [NodeConformance] - test/e2e/common/node/init_container.go:164 - [It] should invoke init containers on a RestartAlways pod [Conformance] - test/e2e/common/node/init_container.go:254 - STEP: creating the pod 12/14/22 16:10:13.919 - Dec 14 16:10:13.920: INFO: PodSpec: initContainers in spec.initContainers - [AfterEach] [sig-node] InitContainer [NodeConformance] + STEP: Creating a kubernetes client 02/25/23 10:12:15.218 + Feb 25 10:12:15.218: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename conformance-tests 02/25/23 10:12:15.221 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:12:15.26 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:12:15.265 + [It] should have at least two untainted nodes [Conformance] + test/e2e/architecture/conformance.go:38 + STEP: Getting node addresses 02/25/23 10:12:15.272 + Feb 25 10:12:15.273: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable + [AfterEach] [sig-architecture] Conformance Tests test/e2e/framework/framework.go:187 - Dec 14 16:10:17.146: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "init-container-6728" for this suite. 12/14/22 16:10:17.153 + Feb 25 10:12:15.287: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "conformance-tests-6894" for this suite. 02/25/23 10:12:15.297 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSS +SSSSSSSS ------------------------------ -[sig-node] Sysctls [LinuxOnly] [NodeConformance] - should reject invalid sysctls [MinimumKubeletVersion:1.21] [Conformance] - test/e2e/common/node/sysctl.go:123 -[BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] - test/e2e/common/node/sysctl.go:37 -[BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] +[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch + watch on custom resource definition objects [Conformance] + test/e2e/apimachinery/crd_watch.go:51 +[BeforeEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:10:17.165 -Dec 14 16:10:17.165: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename sysctl 12/14/22 16:10:17.168 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:10:17.189 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:10:17.192 -[BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] - test/e2e/common/node/sysctl.go:67 -[It] should reject invalid sysctls [MinimumKubeletVersion:1.21] [Conformance] - test/e2e/common/node/sysctl.go:123 -STEP: Creating a pod with one valid and two invalid sysctls 12/14/22 16:10:17.197 -[AfterEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] +STEP: Creating a kubernetes client 02/25/23 10:12:15.319 +Feb 25 10:12:15.319: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename crd-watch 02/25/23 10:12:15.323 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:12:15.362 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:12:15.37 +[It] watch on custom resource definition objects [Conformance] + test/e2e/apimachinery/crd_watch.go:51 +Feb 25 10:12:15.375: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Creating first CR 02/25/23 10:12:18.068 +Feb 25 10:12:18.080: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2023-02-25T10:12:18Z generation:1 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2023-02-25T10:12:18Z]] name:name1 resourceVersion:16278 uid:81e4e2c9-db0e-4810-a391-e41723cafe2d] num:map[num1:9223372036854775807 num2:1000000]]} +STEP: Creating second CR 02/25/23 10:12:28.081 +Feb 25 10:12:28.094: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2023-02-25T10:12:28Z generation:1 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2023-02-25T10:12:28Z]] name:name2 resourceVersion:16314 uid:31de8181-4d6a-4ebe-903e-804279e98608] num:map[num1:9223372036854775807 num2:1000000]]} +STEP: Modifying first CR 02/25/23 10:12:38.094 +Feb 25 10:12:38.108: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2023-02-25T10:12:18Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2023-02-25T10:12:38Z]] name:name1 resourceVersion:16335 uid:81e4e2c9-db0e-4810-a391-e41723cafe2d] num:map[num1:9223372036854775807 num2:1000000]]} +STEP: Modifying second CR 02/25/23 10:12:48.108 +Feb 25 10:12:48.128: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2023-02-25T10:12:28Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2023-02-25T10:12:48Z]] name:name2 resourceVersion:16356 uid:31de8181-4d6a-4ebe-903e-804279e98608] num:map[num1:9223372036854775807 num2:1000000]]} +STEP: Deleting first CR 02/25/23 10:12:58.129 +Feb 25 10:12:58.147: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2023-02-25T10:12:18Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2023-02-25T10:12:38Z]] name:name1 resourceVersion:16377 uid:81e4e2c9-db0e-4810-a391-e41723cafe2d] num:map[num1:9223372036854775807 num2:1000000]]} +STEP: Deleting second CR 02/25/23 10:13:08.148 +Feb 25 10:13:08.160: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2023-02-25T10:12:28Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2023-02-25T10:12:48Z]] name:name2 resourceVersion:16399 uid:31de8181-4d6a-4ebe-903e-804279e98608] num:map[num1:9223372036854775807 num2:1000000]]} +[AfterEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:10:17.209: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "sysctl-7967" for this suite. 12/14/22 16:10:17.219 -{"msg":"PASSED [sig-node] Sysctls [LinuxOnly] [NodeConformance] should reject invalid sysctls [MinimumKubeletVersion:1.21] [Conformance]","completed":97,"skipped":1710,"failed":0} +Feb 25 10:13:18.696: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "crd-watch-7700" for this suite. 02/25/23 10:13:18.706 +{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch watch on custom resource definition objects [Conformance]","completed":99,"skipped":1879,"failed":0} ------------------------------ -• [0.063 seconds] -[sig-node] Sysctls [LinuxOnly] [NodeConformance] -test/e2e/common/node/framework.go:23 - should reject invalid sysctls [MinimumKubeletVersion:1.21] [Conformance] - test/e2e/common/node/sysctl.go:123 +• [SLOW TEST] [63.402 seconds] +[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + CustomResourceDefinition Watch + test/e2e/apimachinery/crd_watch.go:44 + watch on custom resource definition objects [Conformance] + test/e2e/apimachinery/crd_watch.go:51 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] - test/e2e/common/node/sysctl.go:37 - [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] + [BeforeEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:10:17.165 - Dec 14 16:10:17.165: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename sysctl 12/14/22 16:10:17.168 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:10:17.189 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:10:17.192 - [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] - test/e2e/common/node/sysctl.go:67 - [It] should reject invalid sysctls [MinimumKubeletVersion:1.21] [Conformance] - test/e2e/common/node/sysctl.go:123 - STEP: Creating a pod with one valid and two invalid sysctls 12/14/22 16:10:17.197 - [AfterEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] + STEP: Creating a kubernetes client 02/25/23 10:12:15.319 + Feb 25 10:12:15.319: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename crd-watch 02/25/23 10:12:15.323 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:12:15.362 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:12:15.37 + [It] watch on custom resource definition objects [Conformance] + test/e2e/apimachinery/crd_watch.go:51 + Feb 25 10:12:15.375: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Creating first CR 02/25/23 10:12:18.068 + Feb 25 10:12:18.080: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2023-02-25T10:12:18Z generation:1 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2023-02-25T10:12:18Z]] name:name1 resourceVersion:16278 uid:81e4e2c9-db0e-4810-a391-e41723cafe2d] num:map[num1:9223372036854775807 num2:1000000]]} + STEP: Creating second CR 02/25/23 10:12:28.081 + Feb 25 10:12:28.094: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2023-02-25T10:12:28Z generation:1 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2023-02-25T10:12:28Z]] name:name2 resourceVersion:16314 uid:31de8181-4d6a-4ebe-903e-804279e98608] num:map[num1:9223372036854775807 num2:1000000]]} + STEP: Modifying first CR 02/25/23 10:12:38.094 + Feb 25 10:12:38.108: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2023-02-25T10:12:18Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2023-02-25T10:12:38Z]] name:name1 resourceVersion:16335 uid:81e4e2c9-db0e-4810-a391-e41723cafe2d] num:map[num1:9223372036854775807 num2:1000000]]} + STEP: Modifying second CR 02/25/23 10:12:48.108 + Feb 25 10:12:48.128: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2023-02-25T10:12:28Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2023-02-25T10:12:48Z]] name:name2 resourceVersion:16356 uid:31de8181-4d6a-4ebe-903e-804279e98608] num:map[num1:9223372036854775807 num2:1000000]]} + STEP: Deleting first CR 02/25/23 10:12:58.129 + Feb 25 10:12:58.147: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2023-02-25T10:12:18Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2023-02-25T10:12:38Z]] name:name1 resourceVersion:16377 uid:81e4e2c9-db0e-4810-a391-e41723cafe2d] num:map[num1:9223372036854775807 num2:1000000]]} + STEP: Deleting second CR 02/25/23 10:13:08.148 + Feb 25 10:13:08.160: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2023-02-25T10:12:28Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2023-02-25T10:12:48Z]] name:name2 resourceVersion:16399 uid:31de8181-4d6a-4ebe-903e-804279e98608] num:map[num1:9223372036854775807 num2:1000000]]} + [AfterEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:10:17.209: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "sysctl-7967" for this suite. 12/14/22 16:10:17.219 + Feb 25 10:13:18.696: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "crd-watch-7700" for this suite. 02/25/23 10:13:18.706 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSS +SSS ------------------------------ -[sig-cli] Kubectl client Kubectl version - should check is all data is printed [Conformance] - test/e2e/kubectl/kubectl.go:1683 -[BeforeEach] [sig-cli] Kubectl client +[sig-storage] Subpath Atomic writer volumes + should support subpaths with configmap pod with mountPath of existing file [Conformance] + test/e2e/storage/subpath.go:80 +[BeforeEach] [sig-storage] Subpath test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:10:17.233 -Dec 14 16:10:17.233: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 16:10:17.235 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:10:17.261 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:10:17.265 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 -[It] should check is all data is printed [Conformance] - test/e2e/kubectl/kubectl.go:1683 -Dec 14 16:10:17.269: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-1854 version' -Dec 14 16:10:17.384: INFO: stderr: "WARNING: This version information is deprecated and will be replaced with the output from kubectl version --short. Use --output=yaml|json to get the full version.\n" -Dec 14 16:10:17.384: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"25\", GitVersion:\"v1.25.5\", GitCommit:\"804d6167111f6858541cef440ccc53887fbbc96a\", GitTreeState:\"clean\", BuildDate:\"2022-12-08T10:15:02Z\", GoVersion:\"go1.19.4\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nKustomize Version: v4.5.7\nServer Version: version.Info{Major:\"1\", Minor:\"25\", GitVersion:\"v1.25.5\", GitCommit:\"804d6167111f6858541cef440ccc53887fbbc96a\", GitTreeState:\"clean\", BuildDate:\"2022-12-08T10:08:09Z\", GoVersion:\"go1.19.4\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" -[AfterEach] [sig-cli] Kubectl client +STEP: Creating a kubernetes client 02/25/23 10:13:18.722 +Feb 25 10:13:18.723: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename subpath 02/25/23 10:13:18.73 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:13:18.765 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:13:18.769 +[BeforeEach] Atomic writer volumes + test/e2e/storage/subpath.go:40 +STEP: Setting up data 02/25/23 10:13:18.772 +[It] should support subpaths with configmap pod with mountPath of existing file [Conformance] + test/e2e/storage/subpath.go:80 +STEP: Creating pod pod-subpath-test-configmap-8vj6 02/25/23 10:13:18.791 +STEP: Creating a pod to test atomic-volume-subpath 02/25/23 10:13:18.791 +Feb 25 10:13:18.811: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-8vj6" in namespace "subpath-2566" to be "Succeeded or Failed" +Feb 25 10:13:18.823: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Pending", Reason="", readiness=false. Elapsed: 11.005835ms +Feb 25 10:13:20.830: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018255216s +Feb 25 10:13:22.831: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 4.019766933s +Feb 25 10:13:24.833: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 6.021156064s +Feb 25 10:13:26.829: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 8.017230401s +Feb 25 10:13:28.832: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 10.020186158s +Feb 25 10:13:30.833: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 12.021313462s +Feb 25 10:13:32.831: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 14.019517664s +Feb 25 10:13:34.833: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 16.021267803s +Feb 25 10:13:36.830: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 18.017991669s +Feb 25 10:13:38.832: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 20.020342772s +Feb 25 10:13:40.832: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 22.020919948s +Feb 25 10:13:42.830: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=false. Elapsed: 24.018770469s +Feb 25 10:13:44.831: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.019628522s +STEP: Saw pod success 02/25/23 10:13:44.831 +Feb 25 10:13:44.833: INFO: Pod "pod-subpath-test-configmap-8vj6" satisfied condition "Succeeded or Failed" +Feb 25 10:13:44.905: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-subpath-test-configmap-8vj6 container test-container-subpath-configmap-8vj6: +STEP: delete the pod 02/25/23 10:13:45.005 +Feb 25 10:13:45.057: INFO: Waiting for pod pod-subpath-test-configmap-8vj6 to disappear +Feb 25 10:13:45.066: INFO: Pod pod-subpath-test-configmap-8vj6 no longer exists +STEP: Deleting pod pod-subpath-test-configmap-8vj6 02/25/23 10:13:45.066 +Feb 25 10:13:45.066: INFO: Deleting pod "pod-subpath-test-configmap-8vj6" in namespace "subpath-2566" +[AfterEach] [sig-storage] Subpath test/e2e/framework/framework.go:187 -Dec 14 16:10:17.385: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-1854" for this suite. 12/14/22 16:10:17.393 -{"msg":"PASSED [sig-cli] Kubectl client Kubectl version should check is all data is printed [Conformance]","completed":98,"skipped":1725,"failed":0} +Feb 25 10:13:45.083: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "subpath-2566" for this suite. 02/25/23 10:13:45.105 +{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [Conformance]","completed":100,"skipped":1882,"failed":0} ------------------------------ -• [0.173 seconds] -[sig-cli] Kubectl client -test/e2e/kubectl/framework.go:23 - Kubectl version - test/e2e/kubectl/kubectl.go:1677 - should check is all data is printed [Conformance] - test/e2e/kubectl/kubectl.go:1683 +• [SLOW TEST] [26.427 seconds] +[sig-storage] Subpath +test/e2e/storage/utils/framework.go:23 + Atomic writer volumes + test/e2e/storage/subpath.go:36 + should support subpaths with configmap pod with mountPath of existing file [Conformance] + test/e2e/storage/subpath.go:80 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-cli] Kubectl client + [BeforeEach] [sig-storage] Subpath test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:10:17.233 - Dec 14 16:10:17.233: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 16:10:17.235 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:10:17.261 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:10:17.265 - [BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 - [It] should check is all data is printed [Conformance] - test/e2e/kubectl/kubectl.go:1683 - Dec 14 16:10:17.269: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-1854 version' - Dec 14 16:10:17.384: INFO: stderr: "WARNING: This version information is deprecated and will be replaced with the output from kubectl version --short. Use --output=yaml|json to get the full version.\n" - Dec 14 16:10:17.384: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"25\", GitVersion:\"v1.25.5\", GitCommit:\"804d6167111f6858541cef440ccc53887fbbc96a\", GitTreeState:\"clean\", BuildDate:\"2022-12-08T10:15:02Z\", GoVersion:\"go1.19.4\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nKustomize Version: v4.5.7\nServer Version: version.Info{Major:\"1\", Minor:\"25\", GitVersion:\"v1.25.5\", GitCommit:\"804d6167111f6858541cef440ccc53887fbbc96a\", GitTreeState:\"clean\", BuildDate:\"2022-12-08T10:08:09Z\", GoVersion:\"go1.19.4\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" - [AfterEach] [sig-cli] Kubectl client + STEP: Creating a kubernetes client 02/25/23 10:13:18.722 + Feb 25 10:13:18.723: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename subpath 02/25/23 10:13:18.73 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:13:18.765 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:13:18.769 + [BeforeEach] Atomic writer volumes + test/e2e/storage/subpath.go:40 + STEP: Setting up data 02/25/23 10:13:18.772 + [It] should support subpaths with configmap pod with mountPath of existing file [Conformance] + test/e2e/storage/subpath.go:80 + STEP: Creating pod pod-subpath-test-configmap-8vj6 02/25/23 10:13:18.791 + STEP: Creating a pod to test atomic-volume-subpath 02/25/23 10:13:18.791 + Feb 25 10:13:18.811: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-8vj6" in namespace "subpath-2566" to be "Succeeded or Failed" + Feb 25 10:13:18.823: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Pending", Reason="", readiness=false. Elapsed: 11.005835ms + Feb 25 10:13:20.830: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018255216s + Feb 25 10:13:22.831: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 4.019766933s + Feb 25 10:13:24.833: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 6.021156064s + Feb 25 10:13:26.829: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 8.017230401s + Feb 25 10:13:28.832: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 10.020186158s + Feb 25 10:13:30.833: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 12.021313462s + Feb 25 10:13:32.831: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 14.019517664s + Feb 25 10:13:34.833: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 16.021267803s + Feb 25 10:13:36.830: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 18.017991669s + Feb 25 10:13:38.832: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 20.020342772s + Feb 25 10:13:40.832: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=true. Elapsed: 22.020919948s + Feb 25 10:13:42.830: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Running", Reason="", readiness=false. Elapsed: 24.018770469s + Feb 25 10:13:44.831: INFO: Pod "pod-subpath-test-configmap-8vj6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.019628522s + STEP: Saw pod success 02/25/23 10:13:44.831 + Feb 25 10:13:44.833: INFO: Pod "pod-subpath-test-configmap-8vj6" satisfied condition "Succeeded or Failed" + Feb 25 10:13:44.905: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-subpath-test-configmap-8vj6 container test-container-subpath-configmap-8vj6: + STEP: delete the pod 02/25/23 10:13:45.005 + Feb 25 10:13:45.057: INFO: Waiting for pod pod-subpath-test-configmap-8vj6 to disappear + Feb 25 10:13:45.066: INFO: Pod pod-subpath-test-configmap-8vj6 no longer exists + STEP: Deleting pod pod-subpath-test-configmap-8vj6 02/25/23 10:13:45.066 + Feb 25 10:13:45.066: INFO: Deleting pod "pod-subpath-test-configmap-8vj6" in namespace "subpath-2566" + [AfterEach] [sig-storage] Subpath test/e2e/framework/framework.go:187 - Dec 14 16:10:17.385: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-1854" for this suite. 12/14/22 16:10:17.393 + Feb 25 10:13:45.083: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "subpath-2566" for this suite. 02/25/23 10:13:45.105 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-network] Services - should serve a basic endpoint from pods [Conformance] - test/e2e/network/service.go:791 -[BeforeEach] [sig-network] Services +[sig-storage] Subpath Atomic writer volumes + should support subpaths with projected pod [Conformance] + test/e2e/storage/subpath.go:106 +[BeforeEach] [sig-storage] Subpath test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:10:17.408 -Dec 14 16:10:17.408: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 16:10:17.411 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:10:17.436 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:10:17.44 -[BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 -[It] should serve a basic endpoint from pods [Conformance] - test/e2e/network/service.go:791 -STEP: creating service endpoint-test2 in namespace services-2190 12/14/22 16:10:17.444 -STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2190 to expose endpoints map[] 12/14/22 16:10:17.465 -Dec 14 16:10:17.492: INFO: successfully validated that service endpoint-test2 in namespace services-2190 exposes endpoints map[] -STEP: Creating pod pod1 in namespace services-2190 12/14/22 16:10:17.492 -Dec 14 16:10:17.502: INFO: Waiting up to 5m0s for pod "pod1" in namespace "services-2190" to be "running and ready" -Dec 14 16:10:17.511: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 9.218281ms -Dec 14 16:10:17.512: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:10:19.520: INFO: Pod "pod1": Phase="Running", Reason="", readiness=true. Elapsed: 2.018304128s -Dec 14 16:10:19.521: INFO: The phase of Pod pod1 is Running (Ready = true) -Dec 14 16:10:19.521: INFO: Pod "pod1" satisfied condition "running and ready" -STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2190 to expose endpoints map[pod1:[80]] 12/14/22 16:10:19.528 -Dec 14 16:10:19.549: INFO: successfully validated that service endpoint-test2 in namespace services-2190 exposes endpoints map[pod1:[80]] -STEP: Checking if the Service forwards traffic to pod1 12/14/22 16:10:19.549 -Dec 14 16:10:19.550: INFO: Creating new exec pod -Dec 14 16:10:19.557: INFO: Waiting up to 5m0s for pod "execpodxkjnz" in namespace "services-2190" to be "running" -Dec 14 16:10:19.565: INFO: Pod "execpodxkjnz": Phase="Pending", Reason="", readiness=false. Elapsed: 7.506736ms -Dec 14 16:10:21.578: INFO: Pod "execpodxkjnz": Phase="Running", Reason="", readiness=true. Elapsed: 2.020862248s -Dec 14 16:10:21.578: INFO: Pod "execpodxkjnz" satisfied condition "running" -Dec 14 16:10:22.580: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-2190 exec execpodxkjnz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 endpoint-test2 80' -Dec 14 16:10:22.898: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 endpoint-test2 80\nConnection to endpoint-test2 80 port [tcp/http] succeeded!\n" -Dec 14 16:10:22.898: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:10:22.899: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-2190 exec execpodxkjnz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.8.250 80' -Dec 14 16:10:23.104: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.8.250 80\nConnection to 10.233.8.250 80 port [tcp/http] succeeded!\n" -Dec 14 16:10:23.104: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -STEP: Creating pod pod2 in namespace services-2190 12/14/22 16:10:23.104 -Dec 14 16:10:23.113: INFO: Waiting up to 5m0s for pod "pod2" in namespace "services-2190" to be "running and ready" -Dec 14 16:10:23.116: INFO: Pod "pod2": Phase="Pending", Reason="", readiness=false. Elapsed: 3.167629ms -Dec 14 16:10:23.116: INFO: The phase of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:10:25.131: INFO: Pod "pod2": Phase="Running", Reason="", readiness=true. Elapsed: 2.017988371s -Dec 14 16:10:25.131: INFO: The phase of Pod pod2 is Running (Ready = true) -Dec 14 16:10:25.131: INFO: Pod "pod2" satisfied condition "running and ready" -STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2190 to expose endpoints map[pod1:[80] pod2:[80]] 12/14/22 16:10:25.135 -Dec 14 16:10:25.164: INFO: successfully validated that service endpoint-test2 in namespace services-2190 exposes endpoints map[pod1:[80] pod2:[80]] -STEP: Checking if the Service forwards traffic to pod1 and pod2 12/14/22 16:10:25.164 -Dec 14 16:10:26.166: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-2190 exec execpodxkjnz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 endpoint-test2 80' -Dec 14 16:10:26.436: INFO: stderr: "+ nc -v -t -w 2 endpoint-test2 80\n+ echo hostName\nConnection to endpoint-test2 80 port [tcp/http] succeeded!\n" -Dec 14 16:10:26.436: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:10:26.436: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-2190 exec execpodxkjnz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.8.250 80' -Dec 14 16:10:26.707: INFO: stderr: "+ nc -v -t -w 2 10.233.8.250 80\n+ echo hostName\nConnection to 10.233.8.250 80 port [tcp/http] succeeded!\n" -Dec 14 16:10:26.707: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -STEP: Deleting pod pod1 in namespace services-2190 12/14/22 16:10:26.707 -STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2190 to expose endpoints map[pod2:[80]] 12/14/22 16:10:26.747 -Dec 14 16:10:26.771: INFO: successfully validated that service endpoint-test2 in namespace services-2190 exposes endpoints map[pod2:[80]] -STEP: Checking if the Service forwards traffic to pod2 12/14/22 16:10:26.771 -Dec 14 16:10:27.772: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-2190 exec execpodxkjnz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 endpoint-test2 80' -Dec 14 16:10:28.236: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 endpoint-test2 80\nConnection to endpoint-test2 80 port [tcp/http] succeeded!\n" -Dec 14 16:10:28.237: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:10:28.237: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-2190 exec execpodxkjnz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.8.250 80' -Dec 14 16:10:28.581: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.8.250 80\nConnection to 10.233.8.250 80 port [tcp/http] succeeded!\n" -Dec 14 16:10:28.581: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -STEP: Deleting pod pod2 in namespace services-2190 12/14/22 16:10:28.581 -STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2190 to expose endpoints map[] 12/14/22 16:10:28.613 -Dec 14 16:10:29.638: INFO: successfully validated that service endpoint-test2 in namespace services-2190 exposes endpoints map[] -[AfterEach] [sig-network] Services +STEP: Creating a kubernetes client 02/25/23 10:13:45.161 +Feb 25 10:13:45.161: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename subpath 02/25/23 10:13:45.164 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:13:45.245 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:13:45.251 +[BeforeEach] Atomic writer volumes + test/e2e/storage/subpath.go:40 +STEP: Setting up data 02/25/23 10:13:45.257 +[It] should support subpaths with projected pod [Conformance] + test/e2e/storage/subpath.go:106 +STEP: Creating pod pod-subpath-test-projected-8z7r 02/25/23 10:13:45.278 +STEP: Creating a pod to test atomic-volume-subpath 02/25/23 10:13:45.278 +Feb 25 10:13:45.297: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-8z7r" in namespace "subpath-1403" to be "Succeeded or Failed" +Feb 25 10:13:45.304: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Pending", Reason="", readiness=false. Elapsed: 6.47692ms +Feb 25 10:13:47.312: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 2.014866981s +Feb 25 10:13:49.313: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 4.015651707s +Feb 25 10:13:51.312: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 6.014440593s +Feb 25 10:13:53.312: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 8.014865595s +Feb 25 10:13:55.310: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 10.012708154s +Feb 25 10:13:57.314: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 12.017323933s +Feb 25 10:13:59.314: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 14.016648749s +Feb 25 10:14:01.313: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 16.015909829s +Feb 25 10:14:03.316: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 18.018413156s +Feb 25 10:14:05.312: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 20.01509688s +Feb 25 10:14:07.314: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 22.016959429s +Feb 25 10:14:09.313: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=false. Elapsed: 24.016005908s +Feb 25 10:14:11.314: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.017231975s +STEP: Saw pod success 02/25/23 10:14:11.315 +Feb 25 10:14:11.316: INFO: Pod "pod-subpath-test-projected-8z7r" satisfied condition "Succeeded or Failed" +Feb 25 10:14:11.325: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-subpath-test-projected-8z7r container test-container-subpath-projected-8z7r: +STEP: delete the pod 02/25/23 10:14:11.341 +Feb 25 10:14:11.366: INFO: Waiting for pod pod-subpath-test-projected-8z7r to disappear +Feb 25 10:14:11.373: INFO: Pod pod-subpath-test-projected-8z7r no longer exists +STEP: Deleting pod pod-subpath-test-projected-8z7r 02/25/23 10:14:11.373 +Feb 25 10:14:11.374: INFO: Deleting pod "pod-subpath-test-projected-8z7r" in namespace "subpath-1403" +[AfterEach] [sig-storage] Subpath test/e2e/framework/framework.go:187 -Dec 14 16:10:29.675: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-2190" for this suite. 12/14/22 16:10:29.681 -[AfterEach] [sig-network] Services - test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should serve a basic endpoint from pods [Conformance]","completed":99,"skipped":1743,"failed":0} +Feb 25 10:14:11.381: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "subpath-1403" for this suite. 02/25/23 10:14:11.39 +{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [Conformance]","completed":101,"skipped":1915,"failed":0} ------------------------------ -• [SLOW TEST] [12.284 seconds] -[sig-network] Services -test/e2e/network/common/framework.go:23 - should serve a basic endpoint from pods [Conformance] - test/e2e/network/service.go:791 +• [SLOW TEST] [26.243 seconds] +[sig-storage] Subpath +test/e2e/storage/utils/framework.go:23 + Atomic writer volumes + test/e2e/storage/subpath.go:36 + should support subpaths with projected pod [Conformance] + test/e2e/storage/subpath.go:106 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Services + [BeforeEach] [sig-storage] Subpath test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:10:17.408 - Dec 14 16:10:17.408: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 16:10:17.411 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:10:17.436 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:10:17.44 - [BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 - [It] should serve a basic endpoint from pods [Conformance] - test/e2e/network/service.go:791 - STEP: creating service endpoint-test2 in namespace services-2190 12/14/22 16:10:17.444 - STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2190 to expose endpoints map[] 12/14/22 16:10:17.465 - Dec 14 16:10:17.492: INFO: successfully validated that service endpoint-test2 in namespace services-2190 exposes endpoints map[] - STEP: Creating pod pod1 in namespace services-2190 12/14/22 16:10:17.492 - Dec 14 16:10:17.502: INFO: Waiting up to 5m0s for pod "pod1" in namespace "services-2190" to be "running and ready" - Dec 14 16:10:17.511: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 9.218281ms - Dec 14 16:10:17.512: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:10:19.520: INFO: Pod "pod1": Phase="Running", Reason="", readiness=true. Elapsed: 2.018304128s - Dec 14 16:10:19.521: INFO: The phase of Pod pod1 is Running (Ready = true) - Dec 14 16:10:19.521: INFO: Pod "pod1" satisfied condition "running and ready" - STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2190 to expose endpoints map[pod1:[80]] 12/14/22 16:10:19.528 - Dec 14 16:10:19.549: INFO: successfully validated that service endpoint-test2 in namespace services-2190 exposes endpoints map[pod1:[80]] - STEP: Checking if the Service forwards traffic to pod1 12/14/22 16:10:19.549 - Dec 14 16:10:19.550: INFO: Creating new exec pod - Dec 14 16:10:19.557: INFO: Waiting up to 5m0s for pod "execpodxkjnz" in namespace "services-2190" to be "running" - Dec 14 16:10:19.565: INFO: Pod "execpodxkjnz": Phase="Pending", Reason="", readiness=false. Elapsed: 7.506736ms - Dec 14 16:10:21.578: INFO: Pod "execpodxkjnz": Phase="Running", Reason="", readiness=true. Elapsed: 2.020862248s - Dec 14 16:10:21.578: INFO: Pod "execpodxkjnz" satisfied condition "running" - Dec 14 16:10:22.580: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-2190 exec execpodxkjnz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 endpoint-test2 80' - Dec 14 16:10:22.898: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 endpoint-test2 80\nConnection to endpoint-test2 80 port [tcp/http] succeeded!\n" - Dec 14 16:10:22.898: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:10:22.899: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-2190 exec execpodxkjnz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.8.250 80' - Dec 14 16:10:23.104: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.8.250 80\nConnection to 10.233.8.250 80 port [tcp/http] succeeded!\n" - Dec 14 16:10:23.104: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - STEP: Creating pod pod2 in namespace services-2190 12/14/22 16:10:23.104 - Dec 14 16:10:23.113: INFO: Waiting up to 5m0s for pod "pod2" in namespace "services-2190" to be "running and ready" - Dec 14 16:10:23.116: INFO: Pod "pod2": Phase="Pending", Reason="", readiness=false. Elapsed: 3.167629ms - Dec 14 16:10:23.116: INFO: The phase of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:10:25.131: INFO: Pod "pod2": Phase="Running", Reason="", readiness=true. Elapsed: 2.017988371s - Dec 14 16:10:25.131: INFO: The phase of Pod pod2 is Running (Ready = true) - Dec 14 16:10:25.131: INFO: Pod "pod2" satisfied condition "running and ready" - STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2190 to expose endpoints map[pod1:[80] pod2:[80]] 12/14/22 16:10:25.135 - Dec 14 16:10:25.164: INFO: successfully validated that service endpoint-test2 in namespace services-2190 exposes endpoints map[pod1:[80] pod2:[80]] - STEP: Checking if the Service forwards traffic to pod1 and pod2 12/14/22 16:10:25.164 - Dec 14 16:10:26.166: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-2190 exec execpodxkjnz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 endpoint-test2 80' - Dec 14 16:10:26.436: INFO: stderr: "+ nc -v -t -w 2 endpoint-test2 80\n+ echo hostName\nConnection to endpoint-test2 80 port [tcp/http] succeeded!\n" - Dec 14 16:10:26.436: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:10:26.436: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-2190 exec execpodxkjnz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.8.250 80' - Dec 14 16:10:26.707: INFO: stderr: "+ nc -v -t -w 2 10.233.8.250 80\n+ echo hostName\nConnection to 10.233.8.250 80 port [tcp/http] succeeded!\n" - Dec 14 16:10:26.707: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - STEP: Deleting pod pod1 in namespace services-2190 12/14/22 16:10:26.707 - STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2190 to expose endpoints map[pod2:[80]] 12/14/22 16:10:26.747 - Dec 14 16:10:26.771: INFO: successfully validated that service endpoint-test2 in namespace services-2190 exposes endpoints map[pod2:[80]] - STEP: Checking if the Service forwards traffic to pod2 12/14/22 16:10:26.771 - Dec 14 16:10:27.772: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-2190 exec execpodxkjnz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 endpoint-test2 80' - Dec 14 16:10:28.236: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 endpoint-test2 80\nConnection to endpoint-test2 80 port [tcp/http] succeeded!\n" - Dec 14 16:10:28.237: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:10:28.237: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-2190 exec execpodxkjnz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.8.250 80' - Dec 14 16:10:28.581: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.8.250 80\nConnection to 10.233.8.250 80 port [tcp/http] succeeded!\n" - Dec 14 16:10:28.581: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - STEP: Deleting pod pod2 in namespace services-2190 12/14/22 16:10:28.581 - STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2190 to expose endpoints map[] 12/14/22 16:10:28.613 - Dec 14 16:10:29.638: INFO: successfully validated that service endpoint-test2 in namespace services-2190 exposes endpoints map[] - [AfterEach] [sig-network] Services + STEP: Creating a kubernetes client 02/25/23 10:13:45.161 + Feb 25 10:13:45.161: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename subpath 02/25/23 10:13:45.164 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:13:45.245 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:13:45.251 + [BeforeEach] Atomic writer volumes + test/e2e/storage/subpath.go:40 + STEP: Setting up data 02/25/23 10:13:45.257 + [It] should support subpaths with projected pod [Conformance] + test/e2e/storage/subpath.go:106 + STEP: Creating pod pod-subpath-test-projected-8z7r 02/25/23 10:13:45.278 + STEP: Creating a pod to test atomic-volume-subpath 02/25/23 10:13:45.278 + Feb 25 10:13:45.297: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-8z7r" in namespace "subpath-1403" to be "Succeeded or Failed" + Feb 25 10:13:45.304: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Pending", Reason="", readiness=false. Elapsed: 6.47692ms + Feb 25 10:13:47.312: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 2.014866981s + Feb 25 10:13:49.313: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 4.015651707s + Feb 25 10:13:51.312: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 6.014440593s + Feb 25 10:13:53.312: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 8.014865595s + Feb 25 10:13:55.310: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 10.012708154s + Feb 25 10:13:57.314: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 12.017323933s + Feb 25 10:13:59.314: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 14.016648749s + Feb 25 10:14:01.313: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 16.015909829s + Feb 25 10:14:03.316: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 18.018413156s + Feb 25 10:14:05.312: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 20.01509688s + Feb 25 10:14:07.314: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=true. Elapsed: 22.016959429s + Feb 25 10:14:09.313: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Running", Reason="", readiness=false. Elapsed: 24.016005908s + Feb 25 10:14:11.314: INFO: Pod "pod-subpath-test-projected-8z7r": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.017231975s + STEP: Saw pod success 02/25/23 10:14:11.315 + Feb 25 10:14:11.316: INFO: Pod "pod-subpath-test-projected-8z7r" satisfied condition "Succeeded or Failed" + Feb 25 10:14:11.325: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-subpath-test-projected-8z7r container test-container-subpath-projected-8z7r: + STEP: delete the pod 02/25/23 10:14:11.341 + Feb 25 10:14:11.366: INFO: Waiting for pod pod-subpath-test-projected-8z7r to disappear + Feb 25 10:14:11.373: INFO: Pod pod-subpath-test-projected-8z7r no longer exists + STEP: Deleting pod pod-subpath-test-projected-8z7r 02/25/23 10:14:11.373 + Feb 25 10:14:11.374: INFO: Deleting pod "pod-subpath-test-projected-8z7r" in namespace "subpath-1403" + [AfterEach] [sig-storage] Subpath test/e2e/framework/framework.go:187 - Dec 14 16:10:29.675: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-2190" for this suite. 12/14/22 16:10:29.681 - [AfterEach] [sig-network] Services - test/e2e/network/service.go:762 + Feb 25 10:14:11.381: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "subpath-1403" for this suite. 02/25/23 10:14:11.39 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSS +SSSS ------------------------------ -[sig-node] Probing container - should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:180 -[BeforeEach] [sig-node] Probing container +[sig-node] Pods + should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:397 +[BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:10:29.694 -Dec 14 16:10:29.694: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-probe 12/14/22 16:10:29.697 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:10:29.72 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:10:29.723 -[BeforeEach] [sig-node] Probing container - test/e2e/common/node/container_probe.go:59 -[It] should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:180 -STEP: Creating pod liveness-51b1a0d5-050c-40d4-9b15-87496cd29b9c in namespace container-probe-784 12/14/22 16:10:29.729 -Dec 14 16:10:29.741: INFO: Waiting up to 5m0s for pod "liveness-51b1a0d5-050c-40d4-9b15-87496cd29b9c" in namespace "container-probe-784" to be "not pending" -Dec 14 16:10:29.746: INFO: Pod "liveness-51b1a0d5-050c-40d4-9b15-87496cd29b9c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.617289ms -Dec 14 16:10:31.754: INFO: Pod "liveness-51b1a0d5-050c-40d4-9b15-87496cd29b9c": Phase="Running", Reason="", readiness=true. Elapsed: 2.012492894s -Dec 14 16:10:31.754: INFO: Pod "liveness-51b1a0d5-050c-40d4-9b15-87496cd29b9c" satisfied condition "not pending" -Dec 14 16:10:31.754: INFO: Started pod liveness-51b1a0d5-050c-40d4-9b15-87496cd29b9c in namespace container-probe-784 -STEP: checking the pod's current state and verifying that restartCount is present 12/14/22 16:10:31.754 -Dec 14 16:10:31.759: INFO: Initial restart count of pod liveness-51b1a0d5-050c-40d4-9b15-87496cd29b9c is 0 -STEP: deleting the pod 12/14/22 16:14:32.921 -[AfterEach] [sig-node] Probing container +STEP: Creating a kubernetes client 02/25/23 10:14:11.406 +Feb 25 10:14:11.408: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename pods 02/25/23 10:14:11.41 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:14:11.442 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:14:11.448 +[BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 +[It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:397 +STEP: creating the pod 02/25/23 10:14:11.455 +STEP: submitting the pod to kubernetes 02/25/23 10:14:11.455 +Feb 25 10:14:11.470: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759" in namespace "pods-3948" to be "running and ready" +Feb 25 10:14:11.478: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759": Phase="Pending", Reason="", readiness=false. Elapsed: 7.152187ms +Feb 25 10:14:11.478: INFO: The phase of Pod pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:14:13.493: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759": Phase="Running", Reason="", readiness=true. Elapsed: 2.022775575s +Feb 25 10:14:13.493: INFO: The phase of Pod pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759 is Running (Ready = true) +Feb 25 10:14:13.493: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759" satisfied condition "running and ready" +STEP: verifying the pod is in kubernetes 02/25/23 10:14:13.502 +STEP: updating the pod 02/25/23 10:14:13.515 +Feb 25 10:14:14.045: INFO: Successfully updated pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759" +Feb 25 10:14:14.045: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759" in namespace "pods-3948" to be "terminated with reason DeadlineExceeded" +Feb 25 10:14:14.053: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759": Phase="Running", Reason="", readiness=true. Elapsed: 7.546647ms +Feb 25 10:14:16.064: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759": Phase="Running", Reason="", readiness=true. Elapsed: 2.018822987s +Feb 25 10:14:18.061: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759": Phase="Running", Reason="", readiness=false. Elapsed: 4.01568601s +Feb 25 10:14:20.061: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 6.016174365s +Feb 25 10:14:20.062: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759" satisfied condition "terminated with reason DeadlineExceeded" +[AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 -Dec 14 16:14:32.963: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-probe-784" for this suite. 12/14/22 16:14:32.983 -{"msg":"PASSED [sig-node] Probing container should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance]","completed":100,"skipped":1758,"failed":0} +Feb 25 10:14:20.063: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-3948" for this suite. 02/25/23 10:14:20.073 +{"msg":"PASSED [sig-node] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]","completed":102,"skipped":1919,"failed":0} ------------------------------ -• [SLOW TEST] [243.307 seconds] -[sig-node] Probing container +• [SLOW TEST] [8.694 seconds] +[sig-node] Pods test/e2e/common/node/framework.go:23 - should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:180 + should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:397 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Probing container + [BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:10:29.694 - Dec 14 16:10:29.694: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-probe 12/14/22 16:10:29.697 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:10:29.72 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:10:29.723 - [BeforeEach] [sig-node] Probing container - test/e2e/common/node/container_probe.go:59 - [It] should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:180 - STEP: Creating pod liveness-51b1a0d5-050c-40d4-9b15-87496cd29b9c in namespace container-probe-784 12/14/22 16:10:29.729 - Dec 14 16:10:29.741: INFO: Waiting up to 5m0s for pod "liveness-51b1a0d5-050c-40d4-9b15-87496cd29b9c" in namespace "container-probe-784" to be "not pending" - Dec 14 16:10:29.746: INFO: Pod "liveness-51b1a0d5-050c-40d4-9b15-87496cd29b9c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.617289ms - Dec 14 16:10:31.754: INFO: Pod "liveness-51b1a0d5-050c-40d4-9b15-87496cd29b9c": Phase="Running", Reason="", readiness=true. Elapsed: 2.012492894s - Dec 14 16:10:31.754: INFO: Pod "liveness-51b1a0d5-050c-40d4-9b15-87496cd29b9c" satisfied condition "not pending" - Dec 14 16:10:31.754: INFO: Started pod liveness-51b1a0d5-050c-40d4-9b15-87496cd29b9c in namespace container-probe-784 - STEP: checking the pod's current state and verifying that restartCount is present 12/14/22 16:10:31.754 - Dec 14 16:10:31.759: INFO: Initial restart count of pod liveness-51b1a0d5-050c-40d4-9b15-87496cd29b9c is 0 - STEP: deleting the pod 12/14/22 16:14:32.921 - [AfterEach] [sig-node] Probing container + STEP: Creating a kubernetes client 02/25/23 10:14:11.406 + Feb 25 10:14:11.408: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename pods 02/25/23 10:14:11.41 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:14:11.442 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:14:11.448 + [BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 + [It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:397 + STEP: creating the pod 02/25/23 10:14:11.455 + STEP: submitting the pod to kubernetes 02/25/23 10:14:11.455 + Feb 25 10:14:11.470: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759" in namespace "pods-3948" to be "running and ready" + Feb 25 10:14:11.478: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759": Phase="Pending", Reason="", readiness=false. Elapsed: 7.152187ms + Feb 25 10:14:11.478: INFO: The phase of Pod pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:14:13.493: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759": Phase="Running", Reason="", readiness=true. Elapsed: 2.022775575s + Feb 25 10:14:13.493: INFO: The phase of Pod pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759 is Running (Ready = true) + Feb 25 10:14:13.493: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759" satisfied condition "running and ready" + STEP: verifying the pod is in kubernetes 02/25/23 10:14:13.502 + STEP: updating the pod 02/25/23 10:14:13.515 + Feb 25 10:14:14.045: INFO: Successfully updated pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759" + Feb 25 10:14:14.045: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759" in namespace "pods-3948" to be "terminated with reason DeadlineExceeded" + Feb 25 10:14:14.053: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759": Phase="Running", Reason="", readiness=true. Elapsed: 7.546647ms + Feb 25 10:14:16.064: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759": Phase="Running", Reason="", readiness=true. Elapsed: 2.018822987s + Feb 25 10:14:18.061: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759": Phase="Running", Reason="", readiness=false. Elapsed: 4.01568601s + Feb 25 10:14:20.061: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 6.016174365s + Feb 25 10:14:20.062: INFO: Pod "pod-update-activedeadlineseconds-3c26b4da-c0c2-4907-a8ad-82062b33e759" satisfied condition "terminated with reason DeadlineExceeded" + [AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 - Dec 14 16:14:32.963: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-probe-784" for this suite. 12/14/22 16:14:32.983 + Feb 25 10:14:20.063: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "pods-3948" for this suite. 02/25/23 10:14:20.073 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch - watch on custom resource definition objects [Conformance] - test/e2e/apimachinery/crd_watch.go:51 -[BeforeEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] +[sig-storage] ConfigMap + should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:108 +[BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:14:33.004 -Dec 14 16:14:33.004: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename crd-watch 12/14/22 16:14:33.01 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:14:33.037 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:14:33.039 -[It] watch on custom resource definition objects [Conformance] - test/e2e/apimachinery/crd_watch.go:51 -Dec 14 16:14:33.043: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Creating first CR 12/14/22 16:14:35.693 -Dec 14 16:14:35.714: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2022-12-14T16:14:35Z generation:1 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2022-12-14T16:14:35Z]] name:name1 resourceVersion:13991 uid:5258b273-1be6-47e5-b2c6-28a2347d7698] num:map[num1:9223372036854775807 num2:1000000]]} -STEP: Creating second CR 12/14/22 16:14:45.717 -Dec 14 16:14:45.733: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2022-12-14T16:14:45Z generation:1 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2022-12-14T16:14:45Z]] name:name2 resourceVersion:14017 uid:fd8a4f3d-2df8-4847-a867-8705fbafac0b] num:map[num1:9223372036854775807 num2:1000000]]} -STEP: Modifying first CR 12/14/22 16:14:55.734 -Dec 14 16:14:55.749: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2022-12-14T16:14:35Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2022-12-14T16:14:55Z]] name:name1 resourceVersion:14034 uid:5258b273-1be6-47e5-b2c6-28a2347d7698] num:map[num1:9223372036854775807 num2:1000000]]} -STEP: Modifying second CR 12/14/22 16:15:05.75 -Dec 14 16:15:05.769: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2022-12-14T16:14:45Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2022-12-14T16:15:05Z]] name:name2 resourceVersion:14050 uid:fd8a4f3d-2df8-4847-a867-8705fbafac0b] num:map[num1:9223372036854775807 num2:1000000]]} -STEP: Deleting first CR 12/14/22 16:15:15.77 -Dec 14 16:15:15.794: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2022-12-14T16:14:35Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2022-12-14T16:14:55Z]] name:name1 resourceVersion:14066 uid:5258b273-1be6-47e5-b2c6-28a2347d7698] num:map[num1:9223372036854775807 num2:1000000]]} -STEP: Deleting second CR 12/14/22 16:15:25.797 -Dec 14 16:15:25.817: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2022-12-14T16:14:45Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2022-12-14T16:15:05Z]] name:name2 resourceVersion:14081 uid:fd8a4f3d-2df8-4847-a867-8705fbafac0b] num:map[num1:9223372036854775807 num2:1000000]]} -[AfterEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 10:14:20.116 +Feb 25 10:14:20.116: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename configmap 02/25/23 10:14:20.119 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:14:20.149 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:14:20.157 +[It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:108 +STEP: Creating configMap with name configmap-test-volume-map-2e883edc-be0e-41d5-92ca-21a31a11b2bd 02/25/23 10:14:20.164 +STEP: Creating a pod to test consume configMaps 02/25/23 10:14:20.174 +Feb 25 10:14:20.195: INFO: Waiting up to 5m0s for pod "pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88" in namespace "configmap-3939" to be "Succeeded or Failed" +Feb 25 10:14:20.201: INFO: Pod "pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88": Phase="Pending", Reason="", readiness=false. Elapsed: 5.805235ms +Feb 25 10:14:22.210: INFO: Pod "pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015587118s +Feb 25 10:14:24.208: INFO: Pod "pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013484316s +STEP: Saw pod success 02/25/23 10:14:24.208 +Feb 25 10:14:24.209: INFO: Pod "pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88" satisfied condition "Succeeded or Failed" +Feb 25 10:14:24.216: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88 container agnhost-container: +STEP: delete the pod 02/25/23 10:14:24.235 +Feb 25 10:14:24.271: INFO: Waiting for pod pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88 to disappear +Feb 25 10:14:24.279: INFO: Pod pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88 no longer exists +[AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 -Dec 14 16:15:36.352: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-watch-5654" for this suite. 12/14/22 16:15:36.36 -{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] CustomResourceDefinition Watch watch on custom resource definition objects [Conformance]","completed":101,"skipped":1803,"failed":0} +Feb 25 10:14:24.280: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-3939" for this suite. 02/25/23 10:14:24.292 +{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]","completed":103,"skipped":1968,"failed":0} ------------------------------ -• [SLOW TEST] [63.364 seconds] -[sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - CustomResourceDefinition Watch - test/e2e/apimachinery/crd_watch.go:44 - watch on custom resource definition objects [Conformance] - test/e2e/apimachinery/crd_watch.go:51 +• [4.189 seconds] +[sig-storage] ConfigMap +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:108 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] + [BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:14:33.004 - Dec 14 16:14:33.004: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename crd-watch 12/14/22 16:14:33.01 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:14:33.037 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:14:33.039 - [It] watch on custom resource definition objects [Conformance] - test/e2e/apimachinery/crd_watch.go:51 - Dec 14 16:14:33.043: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Creating first CR 12/14/22 16:14:35.693 - Dec 14 16:14:35.714: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2022-12-14T16:14:35Z generation:1 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2022-12-14T16:14:35Z]] name:name1 resourceVersion:13991 uid:5258b273-1be6-47e5-b2c6-28a2347d7698] num:map[num1:9223372036854775807 num2:1000000]]} - STEP: Creating second CR 12/14/22 16:14:45.717 - Dec 14 16:14:45.733: INFO: Got : ADDED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2022-12-14T16:14:45Z generation:1 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2022-12-14T16:14:45Z]] name:name2 resourceVersion:14017 uid:fd8a4f3d-2df8-4847-a867-8705fbafac0b] num:map[num1:9223372036854775807 num2:1000000]]} - STEP: Modifying first CR 12/14/22 16:14:55.734 - Dec 14 16:14:55.749: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2022-12-14T16:14:35Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2022-12-14T16:14:55Z]] name:name1 resourceVersion:14034 uid:5258b273-1be6-47e5-b2c6-28a2347d7698] num:map[num1:9223372036854775807 num2:1000000]]} - STEP: Modifying second CR 12/14/22 16:15:05.75 - Dec 14 16:15:05.769: INFO: Got : MODIFIED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2022-12-14T16:14:45Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2022-12-14T16:15:05Z]] name:name2 resourceVersion:14050 uid:fd8a4f3d-2df8-4847-a867-8705fbafac0b] num:map[num1:9223372036854775807 num2:1000000]]} - STEP: Deleting first CR 12/14/22 16:15:15.77 - Dec 14 16:15:15.794: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2022-12-14T16:14:35Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2022-12-14T16:14:55Z]] name:name1 resourceVersion:14066 uid:5258b273-1be6-47e5-b2c6-28a2347d7698] num:map[num1:9223372036854775807 num2:1000000]]} - STEP: Deleting second CR 12/14/22 16:15:25.797 - Dec 14 16:15:25.817: INFO: Got : DELETED &{map[apiVersion:mygroup.example.com/v1beta1 content:map[key:value] dummy:test kind:WishIHadChosenNoxu metadata:map[creationTimestamp:2022-12-14T16:14:45Z generation:2 managedFields:[map[apiVersion:mygroup.example.com/v1beta1 fieldsType:FieldsV1 fieldsV1:map[f:content:map[.:map[] f:key:map[]] f:dummy:map[] f:num:map[.:map[] f:num1:map[] f:num2:map[]]] manager:e2e.test operation:Update time:2022-12-14T16:15:05Z]] name:name2 resourceVersion:14081 uid:fd8a4f3d-2df8-4847-a867-8705fbafac0b] num:map[num1:9223372036854775807 num2:1000000]]} - [AfterEach] [sig-api-machinery] CustomResourceDefinition Watch [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 10:14:20.116 + Feb 25 10:14:20.116: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename configmap 02/25/23 10:14:20.119 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:14:20.149 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:14:20.157 + [It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:108 + STEP: Creating configMap with name configmap-test-volume-map-2e883edc-be0e-41d5-92ca-21a31a11b2bd 02/25/23 10:14:20.164 + STEP: Creating a pod to test consume configMaps 02/25/23 10:14:20.174 + Feb 25 10:14:20.195: INFO: Waiting up to 5m0s for pod "pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88" in namespace "configmap-3939" to be "Succeeded or Failed" + Feb 25 10:14:20.201: INFO: Pod "pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88": Phase="Pending", Reason="", readiness=false. Elapsed: 5.805235ms + Feb 25 10:14:22.210: INFO: Pod "pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015587118s + Feb 25 10:14:24.208: INFO: Pod "pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013484316s + STEP: Saw pod success 02/25/23 10:14:24.208 + Feb 25 10:14:24.209: INFO: Pod "pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88" satisfied condition "Succeeded or Failed" + Feb 25 10:14:24.216: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88 container agnhost-container: + STEP: delete the pod 02/25/23 10:14:24.235 + Feb 25 10:14:24.271: INFO: Waiting for pod pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88 to disappear + Feb 25 10:14:24.279: INFO: Pod pod-configmaps-cf8020bb-f2cc-4bff-8ff6-aefb24994b88 no longer exists + [AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 - Dec 14 16:15:36.352: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "crd-watch-5654" for this suite. 12/14/22 16:15:36.36 + Feb 25 10:14:24.280: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "configmap-3939" for this suite. 02/25/23 10:14:24.292 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Subpath Atomic writer volumes - should support subpaths with configmap pod [Conformance] - test/e2e/storage/subpath.go:70 -[BeforeEach] [sig-storage] Subpath +[sig-api-machinery] Aggregator + Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] + test/e2e/apimachinery/aggregator.go:100 +[BeforeEach] [sig-api-machinery] Aggregator test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:15:36.372 -Dec 14 16:15:36.372: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename subpath 12/14/22 16:15:36.376 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:15:36.411 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:15:36.414 -[BeforeEach] Atomic writer volumes - test/e2e/storage/subpath.go:40 -STEP: Setting up data 12/14/22 16:15:36.418 -[It] should support subpaths with configmap pod [Conformance] - test/e2e/storage/subpath.go:70 -STEP: Creating pod pod-subpath-test-configmap-mszd 12/14/22 16:15:36.434 -STEP: Creating a pod to test atomic-volume-subpath 12/14/22 16:15:36.434 -Dec 14 16:15:36.455: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-mszd" in namespace "subpath-1612" to be "Succeeded or Failed" -Dec 14 16:15:36.459: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Pending", Reason="", readiness=false. Elapsed: 4.310905ms -Dec 14 16:15:38.469: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 2.013975535s -Dec 14 16:15:40.467: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 4.012105241s -Dec 14 16:15:42.466: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 6.011244304s -Dec 14 16:15:44.467: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 8.012503226s -Dec 14 16:15:46.468: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 10.012762298s -Dec 14 16:15:48.471: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 12.015828118s -Dec 14 16:15:50.471: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 14.015764589s -Dec 14 16:15:52.488: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 16.033674595s -Dec 14 16:15:54.479: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 18.02413345s -Dec 14 16:15:56.469: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 20.014656531s -Dec 14 16:15:58.467: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=false. Elapsed: 22.012749227s -Dec 14 16:16:00.469: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.013828942s -STEP: Saw pod success 12/14/22 16:16:00.469 -Dec 14 16:16:00.469: INFO: Pod "pod-subpath-test-configmap-mszd" satisfied condition "Succeeded or Failed" -Dec 14 16:16:00.475: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-subpath-test-configmap-mszd container test-container-subpath-configmap-mszd: -STEP: delete the pod 12/14/22 16:16:00.502 -Dec 14 16:16:00.526: INFO: Waiting for pod pod-subpath-test-configmap-mszd to disappear -Dec 14 16:16:00.531: INFO: Pod pod-subpath-test-configmap-mszd no longer exists -STEP: Deleting pod pod-subpath-test-configmap-mszd 12/14/22 16:16:00.531 -Dec 14 16:16:00.531: INFO: Deleting pod "pod-subpath-test-configmap-mszd" in namespace "subpath-1612" -[AfterEach] [sig-storage] Subpath +STEP: Creating a kubernetes client 02/25/23 10:14:24.317 +Feb 25 10:14:24.317: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename aggregator 02/25/23 10:14:24.32 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:14:24.359 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:14:24.365 +[BeforeEach] [sig-api-machinery] Aggregator + test/e2e/apimachinery/aggregator.go:78 +Feb 25 10:14:24.372: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +[It] Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] + test/e2e/apimachinery/aggregator.go:100 +STEP: Registering the sample API server. 02/25/23 10:14:24.375 +Feb 25 10:14:25.241: INFO: new replicaset for deployment "sample-apiserver-deployment" is yet to be created +Feb 25 10:14:27.338: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:29.348: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:31.346: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:33.349: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:35.349: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:37.347: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:39.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:41.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:43.346: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:45.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:47.344: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:49.345: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:51.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:53.345: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:55.345: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:57.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:14:59.351: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:01.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:03.348: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:05.347: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:07.352: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:09.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:11.386: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:13.347: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:15.353: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:17.349: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:19.351: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:21.349: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:23.345: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:25.348: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:27.346: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:29.345: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:31.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:33.348: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:35.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:15:37.520: INFO: Waited 155.602786ms for the sample-apiserver to be ready to handle requests. +STEP: Read Status for v1alpha1.wardle.example.com 02/25/23 10:15:37.785 +STEP: kubectl patch apiservice v1alpha1.wardle.example.com -p '{"spec":{"versionPriority": 400}}' 02/25/23 10:15:37.795 +STEP: List APIServices 02/25/23 10:15:37.807 +Feb 25 10:15:37.827: INFO: Found v1alpha1.wardle.example.com in APIServiceList +[AfterEach] [sig-api-machinery] Aggregator + test/e2e/apimachinery/aggregator.go:68 +[AfterEach] [sig-api-machinery] Aggregator test/e2e/framework/framework.go:187 -Dec 14 16:16:00.536: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "subpath-1612" for this suite. 12/14/22 16:16:00.542 -{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [Conformance]","completed":102,"skipped":1851,"failed":0} +Feb 25 10:15:38.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "aggregator-263" for this suite. 02/25/23 10:15:38.255 +{"msg":"PASSED [sig-api-machinery] Aggregator Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance]","completed":104,"skipped":1991,"failed":0} ------------------------------ -• [SLOW TEST] [24.182 seconds] -[sig-storage] Subpath -test/e2e/storage/utils/framework.go:23 - Atomic writer volumes - test/e2e/storage/subpath.go:36 - should support subpaths with configmap pod [Conformance] - test/e2e/storage/subpath.go:70 +• [SLOW TEST] [73.983 seconds] +[sig-api-machinery] Aggregator +test/e2e/apimachinery/framework.go:23 + Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] + test/e2e/apimachinery/aggregator.go:100 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Subpath + [BeforeEach] [sig-api-machinery] Aggregator test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:15:36.372 - Dec 14 16:15:36.372: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename subpath 12/14/22 16:15:36.376 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:15:36.411 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:15:36.414 - [BeforeEach] Atomic writer volumes - test/e2e/storage/subpath.go:40 - STEP: Setting up data 12/14/22 16:15:36.418 - [It] should support subpaths with configmap pod [Conformance] - test/e2e/storage/subpath.go:70 - STEP: Creating pod pod-subpath-test-configmap-mszd 12/14/22 16:15:36.434 - STEP: Creating a pod to test atomic-volume-subpath 12/14/22 16:15:36.434 - Dec 14 16:15:36.455: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-mszd" in namespace "subpath-1612" to be "Succeeded or Failed" - Dec 14 16:15:36.459: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Pending", Reason="", readiness=false. Elapsed: 4.310905ms - Dec 14 16:15:38.469: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 2.013975535s - Dec 14 16:15:40.467: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 4.012105241s - Dec 14 16:15:42.466: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 6.011244304s - Dec 14 16:15:44.467: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 8.012503226s - Dec 14 16:15:46.468: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 10.012762298s - Dec 14 16:15:48.471: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 12.015828118s - Dec 14 16:15:50.471: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 14.015764589s - Dec 14 16:15:52.488: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 16.033674595s - Dec 14 16:15:54.479: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 18.02413345s - Dec 14 16:15:56.469: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=true. Elapsed: 20.014656531s - Dec 14 16:15:58.467: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Running", Reason="", readiness=false. Elapsed: 22.012749227s - Dec 14 16:16:00.469: INFO: Pod "pod-subpath-test-configmap-mszd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.013828942s - STEP: Saw pod success 12/14/22 16:16:00.469 - Dec 14 16:16:00.469: INFO: Pod "pod-subpath-test-configmap-mszd" satisfied condition "Succeeded or Failed" - Dec 14 16:16:00.475: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-subpath-test-configmap-mszd container test-container-subpath-configmap-mszd: - STEP: delete the pod 12/14/22 16:16:00.502 - Dec 14 16:16:00.526: INFO: Waiting for pod pod-subpath-test-configmap-mszd to disappear - Dec 14 16:16:00.531: INFO: Pod pod-subpath-test-configmap-mszd no longer exists - STEP: Deleting pod pod-subpath-test-configmap-mszd 12/14/22 16:16:00.531 - Dec 14 16:16:00.531: INFO: Deleting pod "pod-subpath-test-configmap-mszd" in namespace "subpath-1612" - [AfterEach] [sig-storage] Subpath + STEP: Creating a kubernetes client 02/25/23 10:14:24.317 + Feb 25 10:14:24.317: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename aggregator 02/25/23 10:14:24.32 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:14:24.359 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:14:24.365 + [BeforeEach] [sig-api-machinery] Aggregator + test/e2e/apimachinery/aggregator.go:78 + Feb 25 10:14:24.372: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + [It] Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] + test/e2e/apimachinery/aggregator.go:100 + STEP: Registering the sample API server. 02/25/23 10:14:24.375 + Feb 25 10:14:25.241: INFO: new replicaset for deployment "sample-apiserver-deployment" is yet to be created + Feb 25 10:14:27.338: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:29.348: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:31.346: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:33.349: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:35.349: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:37.347: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:39.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:41.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:43.346: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:45.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:47.344: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:49.345: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:51.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:53.345: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:55.345: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:57.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:14:59.351: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:01.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:03.348: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:05.347: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:07.352: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:09.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:11.386: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:13.347: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:15.353: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:17.349: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:19.351: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:21.349: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:23.345: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:25.348: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:27.346: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:29.345: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:31.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:33.348: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:35.350: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 14, 25, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:15:37.520: INFO: Waited 155.602786ms for the sample-apiserver to be ready to handle requests. + STEP: Read Status for v1alpha1.wardle.example.com 02/25/23 10:15:37.785 + STEP: kubectl patch apiservice v1alpha1.wardle.example.com -p '{"spec":{"versionPriority": 400}}' 02/25/23 10:15:37.795 + STEP: List APIServices 02/25/23 10:15:37.807 + Feb 25 10:15:37.827: INFO: Found v1alpha1.wardle.example.com in APIServiceList + [AfterEach] [sig-api-machinery] Aggregator + test/e2e/apimachinery/aggregator.go:68 + [AfterEach] [sig-api-machinery] Aggregator test/e2e/framework/framework.go:187 - Dec 14 16:16:00.536: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "subpath-1612" for this suite. 12/14/22 16:16:00.542 + Feb 25 10:15:38.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "aggregator-263" for this suite. 02/25/23 10:15:38.255 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSS ------------------------------ -[sig-apps] ReplicationController - should release no longer matching pods [Conformance] - test/e2e/apps/rc.go:100 -[BeforeEach] [sig-apps] ReplicationController +[sig-cli] Kubectl client Kubectl describe + should check if kubectl describe prints relevant information for rc and pods [Conformance] + test/e2e/kubectl/kubectl.go:1274 +[BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:16:00.561 -Dec 14 16:16:00.561: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename replication-controller 12/14/22 16:16:00.564 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:16:00.599 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:16:00.602 -[BeforeEach] [sig-apps] ReplicationController - test/e2e/apps/rc.go:56 -[It] should release no longer matching pods [Conformance] - test/e2e/apps/rc.go:100 -STEP: Given a ReplicationController is created 12/14/22 16:16:00.609 -STEP: When the matched label of one of its pods change 12/14/22 16:16:00.619 -Dec 14 16:16:00.633: INFO: Pod name pod-release: Found 0 pods out of 1 -Dec 14 16:16:05.643: INFO: Pod name pod-release: Found 1 pods out of 1 -STEP: Then the pod is released 12/14/22 16:16:05.66 -[AfterEach] [sig-apps] ReplicationController +STEP: Creating a kubernetes client 02/25/23 10:15:38.314 +Feb 25 10:15:38.314: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 10:15:38.356 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:15:38.391 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:15:38.407 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 +[It] should check if kubectl describe prints relevant information for rc and pods [Conformance] + test/e2e/kubectl/kubectl.go:1274 +Feb 25 10:15:38.417: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-613 create -f -' +Feb 25 10:15:40.936: INFO: stderr: "" +Feb 25 10:15:40.936: INFO: stdout: "replicationcontroller/agnhost-primary created\n" +Feb 25 10:15:40.937: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-613 create -f -' +Feb 25 10:15:41.682: INFO: stderr: "" +Feb 25 10:15:41.682: INFO: stdout: "service/agnhost-primary created\n" +STEP: Waiting for Agnhost primary to start. 02/25/23 10:15:41.682 +Feb 25 10:15:42.700: INFO: Selector matched 1 pods for map[app:agnhost] +Feb 25 10:15:42.701: INFO: Found 0 / 1 +Feb 25 10:15:43.695: INFO: Selector matched 1 pods for map[app:agnhost] +Feb 25 10:15:43.695: INFO: Found 1 / 1 +Feb 25 10:15:43.695: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 +Feb 25 10:15:43.703: INFO: Selector matched 1 pods for map[app:agnhost] +Feb 25 10:15:43.703: INFO: ForEach: Found 1 pods from the filter. Now looping through them. +Feb 25 10:15:43.704: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-613 describe pod agnhost-primary-jvcsz' +Feb 25 10:15:44.115: INFO: stderr: "" +Feb 25 10:15:44.115: INFO: stdout: "Name: agnhost-primary-jvcsz\nNamespace: kubectl-613\nPriority: 0\nService Account: default\nNode: diek3dua9gay-3/192.168.121.44\nStart Time: Sat, 25 Feb 2023 10:15:41 +0000\nLabels: app=agnhost\n role=primary\nAnnotations: \nStatus: Running\nIP: 10.233.66.208\nIPs:\n IP: 10.233.66.208\nControlled By: ReplicationController/agnhost-primary\nContainers:\n agnhost-primary:\n Container ID: cri-o://1319702878f6821912e93e5b00a6b0b9c0f0d7a56e74be685d36adedbc0ba010\n Image: registry.k8s.io/e2e-test-images/agnhost:2.40\n Image ID: registry.k8s.io/e2e-test-images/agnhost@sha256:37dd9ad84b34913b4bc57e00b0a28e8b00f18ca6a5ecec69461aa57402e5c787\n Port: 6379/TCP\n Host Port: 0/TCP\n State: Running\n Started: Sat, 25 Feb 2023 10:15:42 +0000\n Ready: True\n Restart Count: 0\n Environment: \n Mounts:\n /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-t2hd5 (ro)\nConditions:\n Type Status\n Initialized True \n Ready True \n ContainersReady True \n PodScheduled True \nVolumes:\n kube-api-access-t2hd5:\n Type: Projected (a volume that contains injected data from multiple sources)\n TokenExpirationSeconds: 3607\n ConfigMapName: kube-root-ca.crt\n ConfigMapOptional: \n DownwardAPI: true\nQoS Class: BestEffort\nNode-Selectors: \nTolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s\n node.kubernetes.io/unreachable:NoExecute op=Exists for 300s\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal Scheduled 3s default-scheduler Successfully assigned kubectl-613/agnhost-primary-jvcsz to diek3dua9gay-3\n Normal Pulled 2s kubelet Container image \"registry.k8s.io/e2e-test-images/agnhost:2.40\" already present on machine\n Normal Created 2s kubelet Created container agnhost-primary\n Normal Started 2s kubelet Started container agnhost-primary\n" +Feb 25 10:15:44.116: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-613 describe rc agnhost-primary' +Feb 25 10:15:44.498: INFO: stderr: "" +Feb 25 10:15:44.498: INFO: stdout: "Name: agnhost-primary\nNamespace: kubectl-613\nSelector: app=agnhost,role=primary\nLabels: app=agnhost\n role=primary\nAnnotations: \nReplicas: 1 current / 1 desired\nPods Status: 1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n Labels: app=agnhost\n role=primary\n Containers:\n agnhost-primary:\n Image: registry.k8s.io/e2e-test-images/agnhost:2.40\n Port: 6379/TCP\n Host Port: 0/TCP\n Environment: \n Mounts: \n Volumes: \nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal SuccessfulCreate 3s replication-controller Created pod: agnhost-primary-jvcsz\n" +Feb 25 10:15:44.499: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-613 describe service agnhost-primary' +Feb 25 10:15:44.675: INFO: stderr: "" +Feb 25 10:15:44.675: INFO: stdout: "Name: agnhost-primary\nNamespace: kubectl-613\nLabels: app=agnhost\n role=primary\nAnnotations: \nSelector: app=agnhost,role=primary\nType: ClusterIP\nIP Family Policy: SingleStack\nIP Families: IPv4\nIP: 10.233.37.30\nIPs: 10.233.37.30\nPort: 6379/TCP\nTargetPort: agnhost-server/TCP\nEndpoints: 10.233.66.208:6379\nSession Affinity: None\nEvents: \n" +Feb 25 10:15:44.684: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-613 describe node diek3dua9gay-1' +Feb 25 10:15:44.982: INFO: stderr: "" +Feb 25 10:15:44.982: INFO: stdout: "Name: diek3dua9gay-1\nRoles: control-plane\nLabels: beta.kubernetes.io/arch=amd64\n beta.kubernetes.io/os=linux\n kubernetes.io/arch=amd64\n kubernetes.io/hostname=diek3dua9gay-1\n kubernetes.io/os=linux\n node-role.kubernetes.io/control-plane=\n node.kubernetes.io/exclude-from-external-load-balancers=\nAnnotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/crio/crio.sock\n node.alpha.kubernetes.io/ttl: 0\n volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp: Sat, 25 Feb 2023 08:46:21 +0000\nTaints: \nUnschedulable: false\nLease:\n HolderIdentity: diek3dua9gay-1\n AcquireTime: \n RenewTime: Sat, 25 Feb 2023 10:15:41 +0000\nConditions:\n Type Status LastHeartbeatTime LastTransitionTime Reason Message\n ---- ------ ----------------- ------------------ ------ -------\n NetworkUnavailable False Sat, 25 Feb 2023 08:49:07 +0000 Sat, 25 Feb 2023 08:49:07 +0000 CiliumIsUp Cilium is running on this node\n MemoryPressure False Sat, 25 Feb 2023 10:13:26 +0000 Sat, 25 Feb 2023 08:46:12 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available\n DiskPressure False Sat, 25 Feb 2023 10:13:26 +0000 Sat, 25 Feb 2023 08:46:12 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure\n PIDPressure False Sat, 25 Feb 2023 10:13:26 +0000 Sat, 25 Feb 2023 08:46:12 +0000 KubeletHasSufficientPID kubelet has sufficient PID available\n Ready True Sat, 25 Feb 2023 10:13:26 +0000 Sat, 25 Feb 2023 08:49:55 +0000 KubeletReady kubelet is posting ready status. AppArmor enabled\nAddresses:\n InternalIP: 192.168.121.250\n Hostname: diek3dua9gay-1\nCapacity:\n cpu: 2\n ephemeral-storage: 122749536Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 8140212Ki\n pods: 110\nAllocatable:\n cpu: 1600m\n ephemeral-storage: 119410748528\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 3290548Ki\n pods: 110\nSystem Info:\n Machine ID: 31bcfaaed6ce4583a73d198c77e6868b\n System UUID: 31bcfaae-d6ce-4583-a73d-198c77e6868b\n Boot ID: 5f2103e1-3a04-4979-be74-1d6ec37bfc48\n Kernel Version: 5.19.0-32-generic\n OS Image: Ubuntu 22.04.2 LTS\n Operating System: linux\n Architecture: amd64\n Container Runtime Version: cri-o://1.25.2\n Kubelet Version: v1.25.6\n Kube-Proxy Version: v1.25.6\nPodCIDR: 10.233.64.0/24\nPodCIDRs: 10.233.64.0/24\nNon-terminated Pods: (10 in total)\n Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age\n --------- ---- ------------ ---------- --------------- ------------- ---\n cilium-test echo-other-node-cd69fcf6b-hsnk9 0 (0%) 0 (0%) 0 (0%) 0 (0%) 37m\n kube-system cilium-node-init-s84z5 100m (6%) 0 (0%) 100Mi (3%) 0 (0%) 87m\n kube-system cilium-r4lc4 100m (6%) 0 (0%) 100Mi (3%) 0 (0%) 87m\n kube-system coredns-565d847f94-qqvf6 100m (6%) 0 (0%) 70Mi (2%) 170Mi (5%) 86m\n kube-system kube-addon-manager-diek3dua9gay-1 5m (0%) 0 (0%) 50Mi (1%) 0 (0%) 88m\n kube-system kube-apiserver-diek3dua9gay-1 250m (15%) 0 (0%) 0 (0%) 0 (0%) 89m\n kube-system kube-controller-manager-diek3dua9gay-1 200m (12%) 0 (0%) 0 (0%) 0 (0%) 89m\n kube-system kube-proxy-cm4qg 0 (0%) 0 (0%) 0 (0%) 0 (0%) 89m\n kube-system kube-scheduler-diek3dua9gay-1 100m (6%) 0 (0%) 0 (0%) 0 (0%) 89m\n sonobuoy sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-wznrx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 30m\nAllocated resources:\n (Total limits may be over 100 percent, i.e., overcommitted.)\n Resource Requests Limits\n -------- -------- ------\n cpu 855m (53%) 0 (0%)\n memory 320Mi (9%) 170Mi (5%)\n ephemeral-storage 0 (0%) 0 (0%)\n hugepages-1Gi 0 (0%) 0 (0%)\n hugepages-2Mi 0 (0%) 0 (0%)\nEvents: \n" +Feb 25 10:15:44.983: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-613 describe namespace kubectl-613' +Feb 25 10:15:45.227: INFO: stderr: "" +Feb 25 10:15:45.227: INFO: stdout: "Name: kubectl-613\nLabels: e2e-framework=kubectl\n e2e-run=c4f82c74-9faf-42fb-b115-044538789898\n kubernetes.io/metadata.name=kubectl-613\n pod-security.kubernetes.io/enforce=baseline\nAnnotations: \nStatus: Active\n\nNo resource quota.\n\nNo LimitRange resource.\n" +[AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 -Dec 14 16:16:06.676: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "replication-controller-1281" for this suite. 12/14/22 16:16:06.688 -{"msg":"PASSED [sig-apps] ReplicationController should release no longer matching pods [Conformance]","completed":103,"skipped":1890,"failed":0} +Feb 25 10:15:45.228: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-613" for this suite. 02/25/23 10:15:45.236 +{"msg":"PASSED [sig-cli] Kubectl client Kubectl describe should check if kubectl describe prints relevant information for rc and pods [Conformance]","completed":105,"skipped":2001,"failed":0} ------------------------------ -• [SLOW TEST] [6.137 seconds] -[sig-apps] ReplicationController -test/e2e/apps/framework.go:23 - should release no longer matching pods [Conformance] - test/e2e/apps/rc.go:100 +• [SLOW TEST] [6.940 seconds] +[sig-cli] Kubectl client +test/e2e/kubectl/framework.go:23 + Kubectl describe + test/e2e/kubectl/kubectl.go:1268 + should check if kubectl describe prints relevant information for rc and pods [Conformance] + test/e2e/kubectl/kubectl.go:1274 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] ReplicationController + [BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:16:00.561 - Dec 14 16:16:00.561: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename replication-controller 12/14/22 16:16:00.564 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:16:00.599 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:16:00.602 - [BeforeEach] [sig-apps] ReplicationController - test/e2e/apps/rc.go:56 - [It] should release no longer matching pods [Conformance] - test/e2e/apps/rc.go:100 - STEP: Given a ReplicationController is created 12/14/22 16:16:00.609 - STEP: When the matched label of one of its pods change 12/14/22 16:16:00.619 - Dec 14 16:16:00.633: INFO: Pod name pod-release: Found 0 pods out of 1 - Dec 14 16:16:05.643: INFO: Pod name pod-release: Found 1 pods out of 1 - STEP: Then the pod is released 12/14/22 16:16:05.66 - [AfterEach] [sig-apps] ReplicationController + STEP: Creating a kubernetes client 02/25/23 10:15:38.314 + Feb 25 10:15:38.314: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 10:15:38.356 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:15:38.391 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:15:38.407 + [BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 + [It] should check if kubectl describe prints relevant information for rc and pods [Conformance] + test/e2e/kubectl/kubectl.go:1274 + Feb 25 10:15:38.417: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-613 create -f -' + Feb 25 10:15:40.936: INFO: stderr: "" + Feb 25 10:15:40.936: INFO: stdout: "replicationcontroller/agnhost-primary created\n" + Feb 25 10:15:40.937: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-613 create -f -' + Feb 25 10:15:41.682: INFO: stderr: "" + Feb 25 10:15:41.682: INFO: stdout: "service/agnhost-primary created\n" + STEP: Waiting for Agnhost primary to start. 02/25/23 10:15:41.682 + Feb 25 10:15:42.700: INFO: Selector matched 1 pods for map[app:agnhost] + Feb 25 10:15:42.701: INFO: Found 0 / 1 + Feb 25 10:15:43.695: INFO: Selector matched 1 pods for map[app:agnhost] + Feb 25 10:15:43.695: INFO: Found 1 / 1 + Feb 25 10:15:43.695: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 + Feb 25 10:15:43.703: INFO: Selector matched 1 pods for map[app:agnhost] + Feb 25 10:15:43.703: INFO: ForEach: Found 1 pods from the filter. Now looping through them. + Feb 25 10:15:43.704: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-613 describe pod agnhost-primary-jvcsz' + Feb 25 10:15:44.115: INFO: stderr: "" + Feb 25 10:15:44.115: INFO: stdout: "Name: agnhost-primary-jvcsz\nNamespace: kubectl-613\nPriority: 0\nService Account: default\nNode: diek3dua9gay-3/192.168.121.44\nStart Time: Sat, 25 Feb 2023 10:15:41 +0000\nLabels: app=agnhost\n role=primary\nAnnotations: \nStatus: Running\nIP: 10.233.66.208\nIPs:\n IP: 10.233.66.208\nControlled By: ReplicationController/agnhost-primary\nContainers:\n agnhost-primary:\n Container ID: cri-o://1319702878f6821912e93e5b00a6b0b9c0f0d7a56e74be685d36adedbc0ba010\n Image: registry.k8s.io/e2e-test-images/agnhost:2.40\n Image ID: registry.k8s.io/e2e-test-images/agnhost@sha256:37dd9ad84b34913b4bc57e00b0a28e8b00f18ca6a5ecec69461aa57402e5c787\n Port: 6379/TCP\n Host Port: 0/TCP\n State: Running\n Started: Sat, 25 Feb 2023 10:15:42 +0000\n Ready: True\n Restart Count: 0\n Environment: \n Mounts:\n /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-t2hd5 (ro)\nConditions:\n Type Status\n Initialized True \n Ready True \n ContainersReady True \n PodScheduled True \nVolumes:\n kube-api-access-t2hd5:\n Type: Projected (a volume that contains injected data from multiple sources)\n TokenExpirationSeconds: 3607\n ConfigMapName: kube-root-ca.crt\n ConfigMapOptional: \n DownwardAPI: true\nQoS Class: BestEffort\nNode-Selectors: \nTolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s\n node.kubernetes.io/unreachable:NoExecute op=Exists for 300s\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal Scheduled 3s default-scheduler Successfully assigned kubectl-613/agnhost-primary-jvcsz to diek3dua9gay-3\n Normal Pulled 2s kubelet Container image \"registry.k8s.io/e2e-test-images/agnhost:2.40\" already present on machine\n Normal Created 2s kubelet Created container agnhost-primary\n Normal Started 2s kubelet Started container agnhost-primary\n" + Feb 25 10:15:44.116: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-613 describe rc agnhost-primary' + Feb 25 10:15:44.498: INFO: stderr: "" + Feb 25 10:15:44.498: INFO: stdout: "Name: agnhost-primary\nNamespace: kubectl-613\nSelector: app=agnhost,role=primary\nLabels: app=agnhost\n role=primary\nAnnotations: \nReplicas: 1 current / 1 desired\nPods Status: 1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n Labels: app=agnhost\n role=primary\n Containers:\n agnhost-primary:\n Image: registry.k8s.io/e2e-test-images/agnhost:2.40\n Port: 6379/TCP\n Host Port: 0/TCP\n Environment: \n Mounts: \n Volumes: \nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal SuccessfulCreate 3s replication-controller Created pod: agnhost-primary-jvcsz\n" + Feb 25 10:15:44.499: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-613 describe service agnhost-primary' + Feb 25 10:15:44.675: INFO: stderr: "" + Feb 25 10:15:44.675: INFO: stdout: "Name: agnhost-primary\nNamespace: kubectl-613\nLabels: app=agnhost\n role=primary\nAnnotations: \nSelector: app=agnhost,role=primary\nType: ClusterIP\nIP Family Policy: SingleStack\nIP Families: IPv4\nIP: 10.233.37.30\nIPs: 10.233.37.30\nPort: 6379/TCP\nTargetPort: agnhost-server/TCP\nEndpoints: 10.233.66.208:6379\nSession Affinity: None\nEvents: \n" + Feb 25 10:15:44.684: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-613 describe node diek3dua9gay-1' + Feb 25 10:15:44.982: INFO: stderr: "" + Feb 25 10:15:44.982: INFO: stdout: "Name: diek3dua9gay-1\nRoles: control-plane\nLabels: beta.kubernetes.io/arch=amd64\n beta.kubernetes.io/os=linux\n kubernetes.io/arch=amd64\n kubernetes.io/hostname=diek3dua9gay-1\n kubernetes.io/os=linux\n node-role.kubernetes.io/control-plane=\n node.kubernetes.io/exclude-from-external-load-balancers=\nAnnotations: kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/crio/crio.sock\n node.alpha.kubernetes.io/ttl: 0\n volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp: Sat, 25 Feb 2023 08:46:21 +0000\nTaints: \nUnschedulable: false\nLease:\n HolderIdentity: diek3dua9gay-1\n AcquireTime: \n RenewTime: Sat, 25 Feb 2023 10:15:41 +0000\nConditions:\n Type Status LastHeartbeatTime LastTransitionTime Reason Message\n ---- ------ ----------------- ------------------ ------ -------\n NetworkUnavailable False Sat, 25 Feb 2023 08:49:07 +0000 Sat, 25 Feb 2023 08:49:07 +0000 CiliumIsUp Cilium is running on this node\n MemoryPressure False Sat, 25 Feb 2023 10:13:26 +0000 Sat, 25 Feb 2023 08:46:12 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available\n DiskPressure False Sat, 25 Feb 2023 10:13:26 +0000 Sat, 25 Feb 2023 08:46:12 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure\n PIDPressure False Sat, 25 Feb 2023 10:13:26 +0000 Sat, 25 Feb 2023 08:46:12 +0000 KubeletHasSufficientPID kubelet has sufficient PID available\n Ready True Sat, 25 Feb 2023 10:13:26 +0000 Sat, 25 Feb 2023 08:49:55 +0000 KubeletReady kubelet is posting ready status. AppArmor enabled\nAddresses:\n InternalIP: 192.168.121.250\n Hostname: diek3dua9gay-1\nCapacity:\n cpu: 2\n ephemeral-storage: 122749536Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 8140212Ki\n pods: 110\nAllocatable:\n cpu: 1600m\n ephemeral-storage: 119410748528\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 3290548Ki\n pods: 110\nSystem Info:\n Machine ID: 31bcfaaed6ce4583a73d198c77e6868b\n System UUID: 31bcfaae-d6ce-4583-a73d-198c77e6868b\n Boot ID: 5f2103e1-3a04-4979-be74-1d6ec37bfc48\n Kernel Version: 5.19.0-32-generic\n OS Image: Ubuntu 22.04.2 LTS\n Operating System: linux\n Architecture: amd64\n Container Runtime Version: cri-o://1.25.2\n Kubelet Version: v1.25.6\n Kube-Proxy Version: v1.25.6\nPodCIDR: 10.233.64.0/24\nPodCIDRs: 10.233.64.0/24\nNon-terminated Pods: (10 in total)\n Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age\n --------- ---- ------------ ---------- --------------- ------------- ---\n cilium-test echo-other-node-cd69fcf6b-hsnk9 0 (0%) 0 (0%) 0 (0%) 0 (0%) 37m\n kube-system cilium-node-init-s84z5 100m (6%) 0 (0%) 100Mi (3%) 0 (0%) 87m\n kube-system cilium-r4lc4 100m (6%) 0 (0%) 100Mi (3%) 0 (0%) 87m\n kube-system coredns-565d847f94-qqvf6 100m (6%) 0 (0%) 70Mi (2%) 170Mi (5%) 86m\n kube-system kube-addon-manager-diek3dua9gay-1 5m (0%) 0 (0%) 50Mi (1%) 0 (0%) 88m\n kube-system kube-apiserver-diek3dua9gay-1 250m (15%) 0 (0%) 0 (0%) 0 (0%) 89m\n kube-system kube-controller-manager-diek3dua9gay-1 200m (12%) 0 (0%) 0 (0%) 0 (0%) 89m\n kube-system kube-proxy-cm4qg 0 (0%) 0 (0%) 0 (0%) 0 (0%) 89m\n kube-system kube-scheduler-diek3dua9gay-1 100m (6%) 0 (0%) 0 (0%) 0 (0%) 89m\n sonobuoy sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-wznrx 0 (0%) 0 (0%) 0 (0%) 0 (0%) 30m\nAllocated resources:\n (Total limits may be over 100 percent, i.e., overcommitted.)\n Resource Requests Limits\n -------- -------- ------\n cpu 855m (53%) 0 (0%)\n memory 320Mi (9%) 170Mi (5%)\n ephemeral-storage 0 (0%) 0 (0%)\n hugepages-1Gi 0 (0%) 0 (0%)\n hugepages-2Mi 0 (0%) 0 (0%)\nEvents: \n" + Feb 25 10:15:44.983: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-613 describe namespace kubectl-613' + Feb 25 10:15:45.227: INFO: stderr: "" + Feb 25 10:15:45.227: INFO: stdout: "Name: kubectl-613\nLabels: e2e-framework=kubectl\n e2e-run=c4f82c74-9faf-42fb-b115-044538789898\n kubernetes.io/metadata.name=kubectl-613\n pod-security.kubernetes.io/enforce=baseline\nAnnotations: \nStatus: Active\n\nNo resource quota.\n\nNo LimitRange resource.\n" + [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 16:16:06.676: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "replication-controller-1281" for this suite. 12/14/22 16:16:06.688 + Feb 25 10:15:45.228: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-613" for this suite. 02/25/23 10:15:45.236 << End Captured GinkgoWriter Output ------------------------------ -SSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] - should perform rolling updates and roll backs of template modifications [Conformance] - test/e2e/apps/statefulset.go:304 -[BeforeEach] [sig-apps] StatefulSet - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:16:06.7 -Dec 14 16:16:06.700: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename statefulset 12/14/22 16:16:06.704 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:16:06.73 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:16:06.734 -[BeforeEach] [sig-apps] StatefulSet - test/e2e/apps/statefulset.go:96 -[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:111 -STEP: Creating service test in namespace statefulset-7759 12/14/22 16:16:06.739 -[It] should perform rolling updates and roll backs of template modifications [Conformance] - test/e2e/apps/statefulset.go:304 -STEP: Creating a new StatefulSet 12/14/22 16:16:06.747 -Dec 14 16:16:06.762: INFO: Found 0 stateful pods, waiting for 3 -Dec 14 16:16:16.775: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true -Dec 14 16:16:16.776: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true -Dec 14 16:16:16.776: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true -Dec 14 16:16:16.796: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-7759 exec ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' -Dec 14 16:16:17.094: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" -Dec 14 16:16:17.095: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" -Dec 14 16:16:17.095: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - -STEP: Updating StatefulSet template: update image from registry.k8s.io/e2e-test-images/httpd:2.4.38-2 to registry.k8s.io/e2e-test-images/httpd:2.4.39-2 12/14/22 16:16:27.165 -Dec 14 16:16:27.199: INFO: Updating stateful set ss2 -STEP: Creating a new revision 12/14/22 16:16:27.199 -STEP: Updating Pods in reverse ordinal order 12/14/22 16:16:37.231 -Dec 14 16:16:37.237: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-7759 exec ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Dec 14 16:16:37.468: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" -Dec 14 16:16:37.468: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" -Dec 14 16:16:37.468: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - -STEP: Rolling back to a previous revision 12/14/22 16:16:47.511 -Dec 14 16:16:47.511: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-7759 exec ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' -Dec 14 16:16:47.775: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" -Dec 14 16:16:47.775: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" -Dec 14 16:16:47.775: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - -Dec 14 16:16:57.837: INFO: Updating stateful set ss2 -STEP: Rolling back update in reverse ordinal order 12/14/22 16:17:07.891 -Dec 14 16:17:07.896: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-7759 exec ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' -Dec 14 16:17:08.230: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" -Dec 14 16:17:08.230: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" -Dec 14 16:17:08.230: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - -[AfterEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:122 -Dec 14 16:17:18.275: INFO: Deleting all statefulset in ns statefulset-7759 -Dec 14 16:17:18.280: INFO: Scaling statefulset ss2 to 0 -Dec 14 16:17:28.327: INFO: Waiting for statefulset status.replicas updated to 0 -Dec 14 16:17:28.335: INFO: Deleting statefulset ss2 -[AfterEach] [sig-apps] StatefulSet - test/e2e/framework/framework.go:187 -Dec 14 16:17:28.356: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "statefulset-7759" for this suite. 12/14/22 16:17:28.363 -{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance]","completed":104,"skipped":1896,"failed":0} ------------------------------- -• [SLOW TEST] [81.675 seconds] -[sig-apps] StatefulSet -test/e2e/apps/framework.go:23 - Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:101 - should perform rolling updates and roll backs of template modifications [Conformance] - test/e2e/apps/statefulset.go:304 - - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] StatefulSet - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:16:06.7 - Dec 14 16:16:06.700: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename statefulset 12/14/22 16:16:06.704 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:16:06.73 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:16:06.734 - [BeforeEach] [sig-apps] StatefulSet - test/e2e/apps/statefulset.go:96 - [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:111 - STEP: Creating service test in namespace statefulset-7759 12/14/22 16:16:06.739 - [It] should perform rolling updates and roll backs of template modifications [Conformance] - test/e2e/apps/statefulset.go:304 - STEP: Creating a new StatefulSet 12/14/22 16:16:06.747 - Dec 14 16:16:06.762: INFO: Found 0 stateful pods, waiting for 3 - Dec 14 16:16:16.775: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true - Dec 14 16:16:16.776: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true - Dec 14 16:16:16.776: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true - Dec 14 16:16:16.796: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-7759 exec ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' - Dec 14 16:16:17.094: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" - Dec 14 16:16:17.095: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" - Dec 14 16:16:17.095: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - - STEP: Updating StatefulSet template: update image from registry.k8s.io/e2e-test-images/httpd:2.4.38-2 to registry.k8s.io/e2e-test-images/httpd:2.4.39-2 12/14/22 16:16:27.165 - Dec 14 16:16:27.199: INFO: Updating stateful set ss2 - STEP: Creating a new revision 12/14/22 16:16:27.199 - STEP: Updating Pods in reverse ordinal order 12/14/22 16:16:37.231 - Dec 14 16:16:37.237: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-7759 exec ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' - Dec 14 16:16:37.468: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" - Dec 14 16:16:37.468: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" - Dec 14 16:16:37.468: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - - STEP: Rolling back to a previous revision 12/14/22 16:16:47.511 - Dec 14 16:16:47.511: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-7759 exec ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' - Dec 14 16:16:47.775: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" - Dec 14 16:16:47.775: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" - Dec 14 16:16:47.775: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' - - Dec 14 16:16:57.837: INFO: Updating stateful set ss2 - STEP: Rolling back update in reverse ordinal order 12/14/22 16:17:07.891 - Dec 14 16:17:07.896: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=statefulset-7759 exec ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' - Dec 14 16:17:08.230: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" - Dec 14 16:17:08.230: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" - Dec 14 16:17:08.230: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' - - [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:122 - Dec 14 16:17:18.275: INFO: Deleting all statefulset in ns statefulset-7759 - Dec 14 16:17:18.280: INFO: Scaling statefulset ss2 to 0 - Dec 14 16:17:28.327: INFO: Waiting for statefulset status.replicas updated to 0 - Dec 14 16:17:28.335: INFO: Deleting statefulset ss2 - [AfterEach] [sig-apps] StatefulSet - test/e2e/framework/framework.go:187 - Dec 14 16:17:28.356: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "statefulset-7759" for this suite. 12/14/22 16:17:28.363 - << End Captured GinkgoWriter Output ------------------------------- -SSSS ------------------------------- -[sig-api-machinery] Namespaces [Serial] - should apply changes to a namespace status [Conformance] - test/e2e/apimachinery/namespace.go:298 -[BeforeEach] [sig-api-machinery] Namespaces [Serial] +[sig-storage] Secrets + should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:56 +[BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:17:28.376 -Dec 14 16:17:28.376: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename namespaces 12/14/22 16:17:28.383 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:17:28.432 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:17:28.436 -[It] should apply changes to a namespace status [Conformance] - test/e2e/apimachinery/namespace.go:298 -STEP: Read namespace status 12/14/22 16:17:28.44 -Dec 14 16:17:28.447: INFO: Status: v1.NamespaceStatus{Phase:"Active", Conditions:[]v1.NamespaceCondition(nil)} -STEP: Patch namespace status 12/14/22 16:17:28.447 -Dec 14 16:17:28.456: INFO: Status.Condition: v1.NamespaceCondition{Type:"StatusPatch", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Patched by an e2e test"} -STEP: Update namespace status 12/14/22 16:17:28.456 -Dec 14 16:17:28.470: INFO: Status.Condition: v1.NamespaceCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Updated by an e2e test"} -[AfterEach] [sig-api-machinery] Namespaces [Serial] +STEP: Creating a kubernetes client 02/25/23 10:15:45.263 +Feb 25 10:15:45.263: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename secrets 02/25/23 10:15:45.268 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:15:45.302 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:15:45.305 +[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:56 +STEP: Creating secret with name secret-test-d470e3ce-736d-4236-b9a9-9b356ecb7360 02/25/23 10:15:45.311 +STEP: Creating a pod to test consume secrets 02/25/23 10:15:45.322 +Feb 25 10:15:45.338: INFO: Waiting up to 5m0s for pod "pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b" in namespace "secrets-3400" to be "Succeeded or Failed" +Feb 25 10:15:45.343: INFO: Pod "pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.9703ms +Feb 25 10:15:47.352: INFO: Pod "pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014454197s +Feb 25 10:15:49.350: INFO: Pod "pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012445241s +STEP: Saw pod success 02/25/23 10:15:49.35 +Feb 25 10:15:49.351: INFO: Pod "pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b" satisfied condition "Succeeded or Failed" +Feb 25 10:15:49.362: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b container secret-volume-test: +STEP: delete the pod 02/25/23 10:15:49.386 +Feb 25 10:15:49.426: INFO: Waiting for pod pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b to disappear +Feb 25 10:15:49.435: INFO: Pod pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b no longer exists +[AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 -Dec 14 16:17:28.470: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "namespaces-9856" for this suite. 12/14/22 16:17:28.479 -{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should apply changes to a namespace status [Conformance]","completed":105,"skipped":1900,"failed":0} +Feb 25 10:15:49.436: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-3400" for this suite. 02/25/23 10:15:49.449 +{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","completed":106,"skipped":2026,"failed":0} ------------------------------ -• [0.113 seconds] -[sig-api-machinery] Namespaces [Serial] -test/e2e/apimachinery/framework.go:23 - should apply changes to a namespace status [Conformance] - test/e2e/apimachinery/namespace.go:298 +• [4.201 seconds] +[sig-storage] Secrets +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:56 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Namespaces [Serial] + [BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:17:28.376 - Dec 14 16:17:28.376: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename namespaces 12/14/22 16:17:28.383 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:17:28.432 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:17:28.436 - [It] should apply changes to a namespace status [Conformance] - test/e2e/apimachinery/namespace.go:298 - STEP: Read namespace status 12/14/22 16:17:28.44 - Dec 14 16:17:28.447: INFO: Status: v1.NamespaceStatus{Phase:"Active", Conditions:[]v1.NamespaceCondition(nil)} - STEP: Patch namespace status 12/14/22 16:17:28.447 - Dec 14 16:17:28.456: INFO: Status.Condition: v1.NamespaceCondition{Type:"StatusPatch", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Patched by an e2e test"} - STEP: Update namespace status 12/14/22 16:17:28.456 - Dec 14 16:17:28.470: INFO: Status.Condition: v1.NamespaceCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Updated by an e2e test"} - [AfterEach] [sig-api-machinery] Namespaces [Serial] + STEP: Creating a kubernetes client 02/25/23 10:15:45.263 + Feb 25 10:15:45.263: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename secrets 02/25/23 10:15:45.268 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:15:45.302 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:15:45.305 + [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:56 + STEP: Creating secret with name secret-test-d470e3ce-736d-4236-b9a9-9b356ecb7360 02/25/23 10:15:45.311 + STEP: Creating a pod to test consume secrets 02/25/23 10:15:45.322 + Feb 25 10:15:45.338: INFO: Waiting up to 5m0s for pod "pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b" in namespace "secrets-3400" to be "Succeeded or Failed" + Feb 25 10:15:45.343: INFO: Pod "pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.9703ms + Feb 25 10:15:47.352: INFO: Pod "pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014454197s + Feb 25 10:15:49.350: INFO: Pod "pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012445241s + STEP: Saw pod success 02/25/23 10:15:49.35 + Feb 25 10:15:49.351: INFO: Pod "pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b" satisfied condition "Succeeded or Failed" + Feb 25 10:15:49.362: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b container secret-volume-test: + STEP: delete the pod 02/25/23 10:15:49.386 + Feb 25 10:15:49.426: INFO: Waiting for pod pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b to disappear + Feb 25 10:15:49.435: INFO: Pod pod-secrets-726201b7-88e3-4941-af02-e82eedf2ef6b no longer exists + [AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 - Dec 14 16:17:28.470: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "namespaces-9856" for this suite. 12/14/22 16:17:28.479 + Feb 25 10:15:49.436: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "secrets-3400" for this suite. 02/25/23 10:15:49.449 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Variable Expansion - should verify that a failing subpath expansion can be modified during the lifecycle of a container [Slow] [Conformance] - test/e2e/common/node/expansion.go:224 -[BeforeEach] [sig-node] Variable Expansion +[sig-network] DNS + should provide DNS for pods for Subdomain [Conformance] + test/e2e/network/dns.go:290 +[BeforeEach] [sig-network] DNS test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:17:28.492 -Dec 14 16:17:28.492: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename var-expansion 12/14/22 16:17:28.494 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:17:28.518 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:17:28.524 -[It] should verify that a failing subpath expansion can be modified during the lifecycle of a container [Slow] [Conformance] - test/e2e/common/node/expansion.go:224 -STEP: creating the pod with failed condition 12/14/22 16:17:28.527 -Dec 14 16:17:28.541: INFO: Waiting up to 2m0s for pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1" in namespace "var-expansion-7167" to be "running" -Dec 14 16:17:28.548: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 7.134331ms -Dec 14 16:17:30.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018452659s -Dec 14 16:17:32.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 4.018577283s -Dec 14 16:17:34.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 6.015755802s -Dec 14 16:17:36.562: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 8.020315909s -Dec 14 16:17:38.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 10.017424813s -Dec 14 16:17:40.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 12.01794009s -Dec 14 16:17:42.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 14.015258304s -Dec 14 16:17:44.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 16.018142346s -Dec 14 16:17:46.556: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 18.015231619s -Dec 14 16:17:48.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 20.019145459s -Dec 14 16:17:50.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 22.016196071s -Dec 14 16:17:52.563: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 24.021391886s -Dec 14 16:17:54.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 26.018194524s -Dec 14 16:17:56.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 28.017467812s -Dec 14 16:17:58.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 30.017096968s -Dec 14 16:18:00.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 32.017477568s -Dec 14 16:18:02.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 34.017475365s -Dec 14 16:18:04.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 36.01860098s -Dec 14 16:18:06.561: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 38.020076611s -Dec 14 16:18:08.554: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 40.013118666s -Dec 14 16:18:10.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 42.016805864s -Dec 14 16:18:12.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 44.017554393s -Dec 14 16:18:14.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 46.016262661s -Dec 14 16:18:16.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 48.017537952s -Dec 14 16:18:18.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 50.018638566s -Dec 14 16:18:20.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 52.016372059s -Dec 14 16:18:22.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 54.015515701s -Dec 14 16:18:24.561: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 56.01988967s -Dec 14 16:18:26.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 58.016453809s -Dec 14 16:18:28.555: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m0.013246652s -Dec 14 16:18:30.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m2.017004035s -Dec 14 16:18:32.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m4.018911571s -Dec 14 16:18:34.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m6.017650945s -Dec 14 16:18:36.562: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m8.020574561s -Dec 14 16:18:38.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m10.01640985s -Dec 14 16:18:40.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m12.017705435s -Dec 14 16:18:42.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m14.016162179s -Dec 14 16:18:44.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m16.016882497s -Dec 14 16:18:46.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m18.017945585s -Dec 14 16:18:48.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m20.018539881s -Dec 14 16:18:50.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m22.016318575s -Dec 14 16:18:52.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m24.017592878s -Dec 14 16:18:54.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m26.016401494s -Dec 14 16:18:56.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m28.016102116s -Dec 14 16:18:58.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m30.016781298s -Dec 14 16:19:00.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m32.017485792s -Dec 14 16:19:02.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m34.018527007s -Dec 14 16:19:04.561: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m36.020165587s -Dec 14 16:19:06.561: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m38.019925306s -Dec 14 16:19:08.556: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m40.014892224s -Dec 14 16:19:10.563: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m42.021578601s -Dec 14 16:19:12.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m44.019021125s -Dec 14 16:19:14.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m46.015833787s -Dec 14 16:19:16.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m48.017743523s -Dec 14 16:19:18.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m50.015746814s -Dec 14 16:19:20.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m52.017110153s -Dec 14 16:19:22.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m54.017621901s -Dec 14 16:19:24.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m56.017892133s -Dec 14 16:19:26.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m58.018297848s -Dec 14 16:19:28.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 2m0.018087928s -Dec 14 16:19:28.567: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 2m0.025655023s -STEP: updating the pod 12/14/22 16:19:28.567 -Dec 14 16:19:29.109: INFO: Successfully updated pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1" -STEP: waiting for pod running 12/14/22 16:19:29.109 -Dec 14 16:19:29.109: INFO: Waiting up to 2m0s for pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1" in namespace "var-expansion-7167" to be "running" -Dec 14 16:19:29.126: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 16.870391ms -Dec 14 16:19:31.134: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Running", Reason="", readiness=true. Elapsed: 2.024605875s -Dec 14 16:19:31.134: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1" satisfied condition "running" -STEP: deleting the pod gracefully 12/14/22 16:19:31.134 -Dec 14 16:19:31.134: INFO: Deleting pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1" in namespace "var-expansion-7167" -Dec 14 16:19:31.157: INFO: Wait up to 5m0s for pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1" to be fully deleted -[AfterEach] [sig-node] Variable Expansion +STEP: Creating a kubernetes client 02/25/23 10:15:49.467 +Feb 25 10:15:49.468: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename dns 02/25/23 10:15:49.472 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:15:49.519 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:15:49.524 +[It] should provide DNS for pods for Subdomain [Conformance] + test/e2e/network/dns.go:290 +STEP: Creating a test headless service 02/25/23 10:15:49.531 +STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local;sleep 1; done + 02/25/23 10:15:49.544 +STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local;sleep 1; done + 02/25/23 10:15:49.545 +STEP: creating a pod to probe DNS 02/25/23 10:15:49.545 +STEP: submitting the pod to kubernetes 02/25/23 10:15:49.545 +Feb 25 10:15:49.574: INFO: Waiting up to 15m0s for pod "dns-test-ccc69a38-756b-49a2-9726-9383467b6109" in namespace "dns-2923" to be "running" +Feb 25 10:15:49.585: INFO: Pod "dns-test-ccc69a38-756b-49a2-9726-9383467b6109": Phase="Pending", Reason="", readiness=false. Elapsed: 11.167577ms +Feb 25 10:15:51.598: INFO: Pod "dns-test-ccc69a38-756b-49a2-9726-9383467b6109": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023629864s +Feb 25 10:15:53.596: INFO: Pod "dns-test-ccc69a38-756b-49a2-9726-9383467b6109": Phase="Running", Reason="", readiness=true. Elapsed: 4.021849886s +Feb 25 10:15:53.596: INFO: Pod "dns-test-ccc69a38-756b-49a2-9726-9383467b6109" satisfied condition "running" +STEP: retrieving the pod 02/25/23 10:15:53.597 +STEP: looking for the results for each expected name from probers 02/25/23 10:15:53.606 +Feb 25 10:15:53.643: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:53.656: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:53.666: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:53.675: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:53.686: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:53.701: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:53.724: INFO: Unable to read jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:53.736: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:53.736: INFO: Lookups using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local] + +Feb 25 10:15:58.747: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:58.753: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:58.761: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:58.768: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:58.775: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:58.781: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:58.791: INFO: Unable to read jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:58.802: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:15:58.802: INFO: Lookups using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local] + +Feb 25 10:16:03.748: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:03.759: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:03.769: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:03.777: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:03.786: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:03.795: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:03.808: INFO: Unable to read jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:03.826: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:03.826: INFO: Lookups using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local] + +Feb 25 10:16:08.753: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:08.763: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:08.772: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:08.781: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:08.792: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:08.810: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:08.821: INFO: Unable to read jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:08.835: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:08.835: INFO: Lookups using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local] + +Feb 25 10:16:13.765: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:13.774: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:13.784: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:13.794: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:13.805: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:13.822: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:13.833: INFO: Unable to read jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:13.842: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:13.842: INFO: Lookups using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local] + +Feb 25 10:16:18.744: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:18.752: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:18.760: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:18.767: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:18.776: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:18.786: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:18.798: INFO: Unable to read jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:18.809: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:18.809: INFO: Lookups using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local] + +Feb 25 10:16:23.783: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) +Feb 25 10:16:23.844: INFO: Lookups using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 failed for: [wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local] + +Feb 25 10:16:28.788: INFO: DNS probes using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 succeeded + +STEP: deleting the pod 02/25/23 10:16:28.788 +STEP: deleting the test headless service 02/25/23 10:16:28.826 +[AfterEach] [sig-network] DNS test/e2e/framework/framework.go:187 -Dec 14 16:20:03.174: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "var-expansion-7167" for this suite. 12/14/22 16:20:03.182 -{"msg":"PASSED [sig-node] Variable Expansion should verify that a failing subpath expansion can be modified during the lifecycle of a container [Slow] [Conformance]","completed":106,"skipped":1941,"failed":0} +Feb 25 10:16:28.952: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "dns-2923" for this suite. 02/25/23 10:16:29.009 +{"msg":"PASSED [sig-network] DNS should provide DNS for pods for Subdomain [Conformance]","completed":107,"skipped":2047,"failed":0} ------------------------------ -• [SLOW TEST] [154.709 seconds] -[sig-node] Variable Expansion -test/e2e/common/node/framework.go:23 - should verify that a failing subpath expansion can be modified during the lifecycle of a container [Slow] [Conformance] - test/e2e/common/node/expansion.go:224 +• [SLOW TEST] [39.589 seconds] +[sig-network] DNS +test/e2e/network/common/framework.go:23 + should provide DNS for pods for Subdomain [Conformance] + test/e2e/network/dns.go:290 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Variable Expansion + [BeforeEach] [sig-network] DNS test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:17:28.492 - Dec 14 16:17:28.492: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename var-expansion 12/14/22 16:17:28.494 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:17:28.518 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:17:28.524 - [It] should verify that a failing subpath expansion can be modified during the lifecycle of a container [Slow] [Conformance] - test/e2e/common/node/expansion.go:224 - STEP: creating the pod with failed condition 12/14/22 16:17:28.527 - Dec 14 16:17:28.541: INFO: Waiting up to 2m0s for pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1" in namespace "var-expansion-7167" to be "running" - Dec 14 16:17:28.548: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 7.134331ms - Dec 14 16:17:30.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018452659s - Dec 14 16:17:32.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 4.018577283s - Dec 14 16:17:34.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 6.015755802s - Dec 14 16:17:36.562: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 8.020315909s - Dec 14 16:17:38.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 10.017424813s - Dec 14 16:17:40.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 12.01794009s - Dec 14 16:17:42.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 14.015258304s - Dec 14 16:17:44.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 16.018142346s - Dec 14 16:17:46.556: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 18.015231619s - Dec 14 16:17:48.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 20.019145459s - Dec 14 16:17:50.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 22.016196071s - Dec 14 16:17:52.563: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 24.021391886s - Dec 14 16:17:54.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 26.018194524s - Dec 14 16:17:56.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 28.017467812s - Dec 14 16:17:58.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 30.017096968s - Dec 14 16:18:00.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 32.017477568s - Dec 14 16:18:02.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 34.017475365s - Dec 14 16:18:04.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 36.01860098s - Dec 14 16:18:06.561: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 38.020076611s - Dec 14 16:18:08.554: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 40.013118666s - Dec 14 16:18:10.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 42.016805864s - Dec 14 16:18:12.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 44.017554393s - Dec 14 16:18:14.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 46.016262661s - Dec 14 16:18:16.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 48.017537952s - Dec 14 16:18:18.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 50.018638566s - Dec 14 16:18:20.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 52.016372059s - Dec 14 16:18:22.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 54.015515701s - Dec 14 16:18:24.561: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 56.01988967s - Dec 14 16:18:26.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 58.016453809s - Dec 14 16:18:28.555: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m0.013246652s - Dec 14 16:18:30.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m2.017004035s - Dec 14 16:18:32.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m4.018911571s - Dec 14 16:18:34.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m6.017650945s - Dec 14 16:18:36.562: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m8.020574561s - Dec 14 16:18:38.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m10.01640985s - Dec 14 16:18:40.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m12.017705435s - Dec 14 16:18:42.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m14.016162179s - Dec 14 16:18:44.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m16.016882497s - Dec 14 16:18:46.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m18.017945585s - Dec 14 16:18:48.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m20.018539881s - Dec 14 16:18:50.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m22.016318575s - Dec 14 16:18:52.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m24.017592878s - Dec 14 16:18:54.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m26.016401494s - Dec 14 16:18:56.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m28.016102116s - Dec 14 16:18:58.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m30.016781298s - Dec 14 16:19:00.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m32.017485792s - Dec 14 16:19:02.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m34.018527007s - Dec 14 16:19:04.561: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m36.020165587s - Dec 14 16:19:06.561: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m38.019925306s - Dec 14 16:19:08.556: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m40.014892224s - Dec 14 16:19:10.563: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m42.021578601s - Dec 14 16:19:12.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m44.019021125s - Dec 14 16:19:14.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m46.015833787s - Dec 14 16:19:16.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m48.017743523s - Dec 14 16:19:18.557: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m50.015746814s - Dec 14 16:19:20.558: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m52.017110153s - Dec 14 16:19:22.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m54.017621901s - Dec 14 16:19:24.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m56.017892133s - Dec 14 16:19:26.560: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 1m58.018297848s - Dec 14 16:19:28.559: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 2m0.018087928s - Dec 14 16:19:28.567: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 2m0.025655023s - STEP: updating the pod 12/14/22 16:19:28.567 - Dec 14 16:19:29.109: INFO: Successfully updated pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1" - STEP: waiting for pod running 12/14/22 16:19:29.109 - Dec 14 16:19:29.109: INFO: Waiting up to 2m0s for pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1" in namespace "var-expansion-7167" to be "running" - Dec 14 16:19:29.126: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Pending", Reason="", readiness=false. Elapsed: 16.870391ms - Dec 14 16:19:31.134: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1": Phase="Running", Reason="", readiness=true. Elapsed: 2.024605875s - Dec 14 16:19:31.134: INFO: Pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1" satisfied condition "running" - STEP: deleting the pod gracefully 12/14/22 16:19:31.134 - Dec 14 16:19:31.134: INFO: Deleting pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1" in namespace "var-expansion-7167" - Dec 14 16:19:31.157: INFO: Wait up to 5m0s for pod "var-expansion-7c5434de-b80f-42bb-838d-bfd9060481d1" to be fully deleted - [AfterEach] [sig-node] Variable Expansion + STEP: Creating a kubernetes client 02/25/23 10:15:49.467 + Feb 25 10:15:49.468: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename dns 02/25/23 10:15:49.472 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:15:49.519 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:15:49.524 + [It] should provide DNS for pods for Subdomain [Conformance] + test/e2e/network/dns.go:290 + STEP: Creating a test headless service 02/25/23 10:15:49.531 + STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local;sleep 1; done + 02/25/23 10:15:49.544 + STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-2923.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local;sleep 1; done + 02/25/23 10:15:49.545 + STEP: creating a pod to probe DNS 02/25/23 10:15:49.545 + STEP: submitting the pod to kubernetes 02/25/23 10:15:49.545 + Feb 25 10:15:49.574: INFO: Waiting up to 15m0s for pod "dns-test-ccc69a38-756b-49a2-9726-9383467b6109" in namespace "dns-2923" to be "running" + Feb 25 10:15:49.585: INFO: Pod "dns-test-ccc69a38-756b-49a2-9726-9383467b6109": Phase="Pending", Reason="", readiness=false. Elapsed: 11.167577ms + Feb 25 10:15:51.598: INFO: Pod "dns-test-ccc69a38-756b-49a2-9726-9383467b6109": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023629864s + Feb 25 10:15:53.596: INFO: Pod "dns-test-ccc69a38-756b-49a2-9726-9383467b6109": Phase="Running", Reason="", readiness=true. Elapsed: 4.021849886s + Feb 25 10:15:53.596: INFO: Pod "dns-test-ccc69a38-756b-49a2-9726-9383467b6109" satisfied condition "running" + STEP: retrieving the pod 02/25/23 10:15:53.597 + STEP: looking for the results for each expected name from probers 02/25/23 10:15:53.606 + Feb 25 10:15:53.643: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:53.656: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:53.666: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:53.675: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:53.686: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:53.701: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:53.724: INFO: Unable to read jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:53.736: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:53.736: INFO: Lookups using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local] + + Feb 25 10:15:58.747: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:58.753: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:58.761: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:58.768: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:58.775: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:58.781: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:58.791: INFO: Unable to read jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:58.802: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:15:58.802: INFO: Lookups using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local] + + Feb 25 10:16:03.748: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:03.759: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:03.769: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:03.777: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:03.786: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:03.795: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:03.808: INFO: Unable to read jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:03.826: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:03.826: INFO: Lookups using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local] + + Feb 25 10:16:08.753: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:08.763: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:08.772: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:08.781: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:08.792: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:08.810: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:08.821: INFO: Unable to read jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:08.835: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:08.835: INFO: Lookups using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local] + + Feb 25 10:16:13.765: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:13.774: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:13.784: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:13.794: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:13.805: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:13.822: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:13.833: INFO: Unable to read jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:13.842: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:13.842: INFO: Lookups using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local] + + Feb 25 10:16:18.744: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:18.752: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:18.760: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:18.767: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:18.776: INFO: Unable to read jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:18.786: INFO: Unable to read jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:18.798: INFO: Unable to read jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:18.809: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:18.809: INFO: Lookups using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-querier-2.dns-test-service-2.dns-2923.svc.cluster.local jessie_udp@dns-test-service-2.dns-2923.svc.cluster.local jessie_tcp@dns-test-service-2.dns-2923.svc.cluster.local] + + Feb 25 10:16:23.783: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local from pod dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109: the server could not find the requested resource (get pods dns-test-ccc69a38-756b-49a2-9726-9383467b6109) + Feb 25 10:16:23.844: INFO: Lookups using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 failed for: [wheezy_udp@dns-test-service-2.dns-2923.svc.cluster.local] + + Feb 25 10:16:28.788: INFO: DNS probes using dns-2923/dns-test-ccc69a38-756b-49a2-9726-9383467b6109 succeeded + + STEP: deleting the pod 02/25/23 10:16:28.788 + STEP: deleting the test headless service 02/25/23 10:16:28.826 + [AfterEach] [sig-network] DNS test/e2e/framework/framework.go:187 - Dec 14 16:20:03.174: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "var-expansion-7167" for this suite. 12/14/22 16:20:03.182 + Feb 25 10:16:28.952: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "dns-2923" for this suite. 02/25/23 10:16:29.009 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSS +S ------------------------------ -[sig-storage] CSIStorageCapacity - should support CSIStorageCapacities API operations [Conformance] - test/e2e/storage/csistoragecapacity.go:49 -[BeforeEach] [sig-storage] CSIStorageCapacity +[sig-cli] Kubectl client Update Demo + should create and stop a replication controller [Conformance] + test/e2e/kubectl/kubectl.go:337 +[BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:20:03.204 -Dec 14 16:20:03.204: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename csistoragecapacity 12/14/22 16:20:03.209 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:20:03.245 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:20:03.249 -[It] should support CSIStorageCapacities API operations [Conformance] - test/e2e/storage/csistoragecapacity.go:49 -STEP: getting /apis 12/14/22 16:20:03.252 -STEP: getting /apis/storage.k8s.io 12/14/22 16:20:03.254 -STEP: getting /apis/storage.k8s.io/v1 12/14/22 16:20:03.256 -STEP: creating 12/14/22 16:20:03.257 -STEP: watching 12/14/22 16:20:03.279 -Dec 14 16:20:03.279: INFO: starting watch -STEP: getting 12/14/22 16:20:03.287 -STEP: listing in namespace 12/14/22 16:20:03.291 -STEP: listing across namespaces 12/14/22 16:20:03.295 -STEP: patching 12/14/22 16:20:03.299 -STEP: updating 12/14/22 16:20:03.305 -Dec 14 16:20:03.315: INFO: waiting for watch events with expected annotations in namespace -Dec 14 16:20:03.315: INFO: waiting for watch events with expected annotations across namespace -STEP: deleting 12/14/22 16:20:03.315 -STEP: deleting a collection 12/14/22 16:20:03.334 -[AfterEach] [sig-storage] CSIStorageCapacity - test/e2e/framework/framework.go:187 -Dec 14 16:20:03.361: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "csistoragecapacity-1151" for this suite. 12/14/22 16:20:03.368 -{"msg":"PASSED [sig-storage] CSIStorageCapacity should support CSIStorageCapacities API operations [Conformance]","completed":107,"skipped":1953,"failed":0} ------------------------------- -• [0.173 seconds] -[sig-storage] CSIStorageCapacity -test/e2e/storage/utils/framework.go:23 - should support CSIStorageCapacities API operations [Conformance] - test/e2e/storage/csistoragecapacity.go:49 +STEP: Creating a kubernetes client 02/25/23 10:16:29.057 +Feb 25 10:16:29.057: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 10:16:29.065 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:16:29.223 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:16:29.231 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 +[BeforeEach] Update Demo + test/e2e/kubectl/kubectl.go:324 +[It] should create and stop a replication controller [Conformance] + test/e2e/kubectl/kubectl.go:337 +STEP: creating a replication controller 02/25/23 10:16:29.24 +Feb 25 10:16:29.241: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 create -f -' +Feb 25 10:16:30.020: INFO: stderr: "" +Feb 25 10:16:30.020: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" +STEP: waiting for all containers in name=update-demo pods to come up. 02/25/23 10:16:30.02 +Feb 25 10:16:30.022: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' +Feb 25 10:16:30.269: INFO: stderr: "" +Feb 25 10:16:30.269: INFO: stdout: "update-demo-nautilus-8j946 update-demo-nautilus-n9qzd " +Feb 25 10:16:30.275: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-8j946 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' +Feb 25 10:16:30.632: INFO: stderr: "" +Feb 25 10:16:30.632: INFO: stdout: "" +Feb 25 10:16:30.632: INFO: update-demo-nautilus-8j946 is created but not running +Feb 25 10:16:35.643: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' +Feb 25 10:16:35.946: INFO: stderr: "" +Feb 25 10:16:35.946: INFO: stdout: "update-demo-nautilus-8j946 update-demo-nautilus-n9qzd " +Feb 25 10:16:35.946: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-8j946 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' +Feb 25 10:16:36.270: INFO: stderr: "" +Feb 25 10:16:36.270: INFO: stdout: "" +Feb 25 10:16:36.270: INFO: update-demo-nautilus-8j946 is created but not running +Feb 25 10:16:41.272: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' +Feb 25 10:16:41.528: INFO: stderr: "" +Feb 25 10:16:41.529: INFO: stdout: "update-demo-nautilus-8j946 update-demo-nautilus-n9qzd " +Feb 25 10:16:41.529: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-8j946 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' +Feb 25 10:16:41.806: INFO: stderr: "" +Feb 25 10:16:41.806: INFO: stdout: "" +Feb 25 10:16:41.806: INFO: update-demo-nautilus-8j946 is created but not running +Feb 25 10:16:46.808: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' +Feb 25 10:16:47.045: INFO: stderr: "" +Feb 25 10:16:47.045: INFO: stdout: "update-demo-nautilus-8j946 update-demo-nautilus-n9qzd " +Feb 25 10:16:47.047: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-8j946 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' +Feb 25 10:16:47.293: INFO: stderr: "" +Feb 25 10:16:47.293: INFO: stdout: "" +Feb 25 10:16:47.293: INFO: update-demo-nautilus-8j946 is created but not running +Feb 25 10:16:52.294: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' +Feb 25 10:16:52.484: INFO: stderr: "" +Feb 25 10:16:52.484: INFO: stdout: "update-demo-nautilus-8j946 update-demo-nautilus-n9qzd " +Feb 25 10:16:52.484: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-8j946 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' +Feb 25 10:16:52.628: INFO: stderr: "" +Feb 25 10:16:52.628: INFO: stdout: "true" +Feb 25 10:16:52.628: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-8j946 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' +Feb 25 10:16:52.780: INFO: stderr: "" +Feb 25 10:16:52.780: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" +Feb 25 10:16:52.780: INFO: validating pod update-demo-nautilus-8j946 +Feb 25 10:16:52.797: INFO: got data: { + "image": "nautilus.jpg" +} - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] CSIStorageCapacity - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:20:03.204 - Dec 14 16:20:03.204: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename csistoragecapacity 12/14/22 16:20:03.209 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:20:03.245 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:20:03.249 - [It] should support CSIStorageCapacities API operations [Conformance] - test/e2e/storage/csistoragecapacity.go:49 - STEP: getting /apis 12/14/22 16:20:03.252 - STEP: getting /apis/storage.k8s.io 12/14/22 16:20:03.254 - STEP: getting /apis/storage.k8s.io/v1 12/14/22 16:20:03.256 - STEP: creating 12/14/22 16:20:03.257 - STEP: watching 12/14/22 16:20:03.279 - Dec 14 16:20:03.279: INFO: starting watch - STEP: getting 12/14/22 16:20:03.287 - STEP: listing in namespace 12/14/22 16:20:03.291 - STEP: listing across namespaces 12/14/22 16:20:03.295 - STEP: patching 12/14/22 16:20:03.299 - STEP: updating 12/14/22 16:20:03.305 - Dec 14 16:20:03.315: INFO: waiting for watch events with expected annotations in namespace - Dec 14 16:20:03.315: INFO: waiting for watch events with expected annotations across namespace - STEP: deleting 12/14/22 16:20:03.315 - STEP: deleting a collection 12/14/22 16:20:03.334 - [AfterEach] [sig-storage] CSIStorageCapacity - test/e2e/framework/framework.go:187 - Dec 14 16:20:03.361: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "csistoragecapacity-1151" for this suite. 12/14/22 16:20:03.368 - << End Captured GinkgoWriter Output ------------------------------- -SSSS ------------------------------- -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - works for CRD with validation schema [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:68 -[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:20:03.381 -Dec 14 16:20:03.381: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 16:20:03.383 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:20:03.415 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:20:03.419 -[It] works for CRD with validation schema [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:68 -Dec 14 16:20:03.424: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: kubectl validation (kubectl create and apply) allows request with known and required properties 12/14/22 16:20:06.524 -Dec 14 16:20:06.526: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 create -f -' -Dec 14 16:20:07.988: INFO: stderr: "" -Dec 14 16:20:07.988: INFO: stdout: "e2e-test-crd-publish-openapi-3117-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" -Dec 14 16:20:07.988: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 delete e2e-test-crd-publish-openapi-3117-crds test-foo' -Dec 14 16:20:08.272: INFO: stderr: "" -Dec 14 16:20:08.272: INFO: stdout: "e2e-test-crd-publish-openapi-3117-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" -Dec 14 16:20:08.273: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 apply -f -' -Dec 14 16:20:08.775: INFO: stderr: "" -Dec 14 16:20:08.775: INFO: stdout: "e2e-test-crd-publish-openapi-3117-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" -Dec 14 16:20:08.775: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 delete e2e-test-crd-publish-openapi-3117-crds test-foo' -Dec 14 16:20:08.908: INFO: stderr: "" -Dec 14 16:20:08.908: INFO: stdout: "e2e-test-crd-publish-openapi-3117-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" -STEP: kubectl validation (kubectl create and apply) rejects request with value outside defined enum values 12/14/22 16:20:08.908 -Dec 14 16:20:08.909: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 create -f -' -Dec 14 16:20:09.711: INFO: rc: 1 -STEP: kubectl validation (kubectl create and apply) rejects request with unknown properties when disallowed by the schema 12/14/22 16:20:09.711 -Dec 14 16:20:09.711: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 create -f -' -Dec 14 16:20:10.136: INFO: rc: 1 -Dec 14 16:20:10.137: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 apply -f -' -Dec 14 16:20:10.659: INFO: rc: 1 -STEP: kubectl validation (kubectl create and apply) rejects request without required properties 12/14/22 16:20:10.659 -Dec 14 16:20:10.660: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 create -f -' -Dec 14 16:20:11.008: INFO: rc: 1 -Dec 14 16:20:11.008: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 apply -f -' -Dec 14 16:20:11.323: INFO: rc: 1 -STEP: kubectl explain works to explain CR properties 12/14/22 16:20:11.323 -Dec 14 16:20:11.324: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 explain e2e-test-crd-publish-openapi-3117-crds' -Dec 14 16:20:11.673: INFO: stderr: "" -Dec 14 16:20:11.673: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-3117-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nDESCRIPTION:\n Foo CRD for Testing\n\nFIELDS:\n apiVersion\t\n APIVersion defines the versioned schema of this representation of an\n object. Servers should convert recognized schemas to the latest internal\n value, and may reject unrecognized values. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n kind\t\n Kind is a string value representing the REST resource this object\n represents. Servers may infer this from the endpoint the client submits\n requests to. Cannot be updated. In CamelCase. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n metadata\t\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n spec\t\n Specification of Foo\n\n status\t\n Status of Foo\n\n" -STEP: kubectl explain works to explain CR properties recursively 12/14/22 16:20:11.674 -Dec 14 16:20:11.675: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 explain e2e-test-crd-publish-openapi-3117-crds.metadata' -Dec 14 16:20:12.231: INFO: stderr: "" -Dec 14 16:20:12.231: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-3117-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: metadata \n\nDESCRIPTION:\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n ObjectMeta is metadata that all persisted resources must have, which\n includes all objects users must create.\n\nFIELDS:\n annotations\t\n Annotations is an unstructured key value map stored with a resource that\n may be set by external tools to store and retrieve arbitrary metadata. They\n are not queryable and should be preserved when modifying objects. More\n info: http://kubernetes.io/docs/user-guide/annotations\n\n creationTimestamp\t\n CreationTimestamp is a timestamp representing the server time when this\n object was created. It is not guaranteed to be set in happens-before order\n across separate operations. Clients may not set this value. It is\n represented in RFC3339 form and is in UTC.\n\n Populated by the system. Read-only. Null for lists. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n deletionGracePeriodSeconds\t\n Number of seconds allowed for this object to gracefully terminate before it\n will be removed from the system. Only set when deletionTimestamp is also\n set. May only be shortened. Read-only.\n\n deletionTimestamp\t\n DeletionTimestamp is RFC 3339 date and time at which this resource will be\n deleted. This field is set by the server when a graceful deletion is\n requested by the user, and is not directly settable by a client. The\n resource is expected to be deleted (no longer visible from resource lists,\n and not reachable by name) after the time in this field, once the\n finalizers list is empty. As long as the finalizers list contains items,\n deletion is blocked. Once the deletionTimestamp is set, this value may not\n be unset or be set further into the future, although it may be shortened or\n the resource may be deleted prior to this time. For example, a user may\n request that a pod is deleted in 30 seconds. The Kubelet will react by\n sending a graceful termination signal to the containers in the pod. After\n that 30 seconds, the Kubelet will send a hard termination signal (SIGKILL)\n to the container and after cleanup, remove the pod from the API. In the\n presence of network partitions, this object may still exist after this\n timestamp, until an administrator or automated process can determine the\n resource is fully terminated. If not set, graceful deletion of the object\n has not been requested.\n\n Populated by the system when a graceful deletion is requested. Read-only.\n More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n finalizers\t<[]string>\n Must be empty before the object is deleted from the registry. Each entry is\n an identifier for the responsible component that will remove the entry from\n the list. If the deletionTimestamp of the object is non-nil, entries in\n this list can only be removed. Finalizers may be processed and removed in\n any order. Order is NOT enforced because it introduces significant risk of\n stuck finalizers. finalizers is a shared field, any actor with permission\n can reorder it. If the finalizer list is processed in order, then this can\n lead to a situation in which the component responsible for the first\n finalizer in the list is waiting for a signal (field value, external\n system, or other) produced by a component responsible for a finalizer later\n in the list, resulting in a deadlock. Without enforced ordering finalizers\n are free to order amongst themselves and are not vulnerable to ordering\n changes in the list.\n\n generateName\t\n GenerateName is an optional prefix, used by the server, to generate a\n unique name ONLY IF the Name field has not been provided. If this field is\n used, the name returned to the client will be different than the name\n passed. This value will also be combined with a unique suffix. The provided\n value has the same validation rules as the Name field, and may be truncated\n by the length of the suffix required to make the value unique on the\n server.\n\n If this field is specified and the generated name exists, the server will\n return a 409.\n\n Applied only if Name is not specified. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency\n\n generation\t\n A sequence number representing a specific generation of the desired state.\n Populated by the system. Read-only.\n\n labels\t\n Map of string keys and values that can be used to organize and categorize\n (scope and select) objects. May match selectors of replication controllers\n and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n managedFields\t<[]Object>\n ManagedFields maps workflow-id and version to the set of fields that are\n managed by that workflow. This is mostly for internal housekeeping, and\n users typically shouldn't need to set or understand this field. A workflow\n can be the user's name, a controller's name, or the name of a specific\n apply path like \"ci-cd\". The set of fields is always in the version that\n the workflow used when modifying the object.\n\n name\t\n Name must be unique within a namespace. Is required when creating\n resources, although some resources may allow a client to request the\n generation of an appropriate name automatically. Name is primarily intended\n for creation idempotence and configuration definition. Cannot be updated.\n More info: http://kubernetes.io/docs/user-guide/identifiers#names\n\n namespace\t\n Namespace defines the space within which each name must be unique. An empty\n namespace is equivalent to the \"default\" namespace, but \"default\" is the\n canonical representation. Not all objects are required to be scoped to a\n namespace - the value of this field for those objects will be empty.\n\n Must be a DNS_LABEL. Cannot be updated. More info:\n http://kubernetes.io/docs/user-guide/namespaces\n\n ownerReferences\t<[]Object>\n List of objects depended by this object. If ALL objects in the list have\n been deleted, this object will be garbage collected. If this object is\n managed by a controller, then an entry in this list will point to this\n controller, with the controller field set to true. There cannot be more\n than one managing controller.\n\n resourceVersion\t\n An opaque value that represents the internal version of this object that\n can be used by clients to determine when objects have changed. May be used\n for optimistic concurrency, change detection, and the watch operation on a\n resource or set of resources. Clients must treat these values as opaque and\n passed unmodified back to the server. They may only be valid for a\n particular resource or set of resources.\n\n Populated by the system. Read-only. Value must be treated as opaque by\n clients and . More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency\n\n selfLink\t\n Deprecated: selfLink is a legacy read-only field that is no longer\n populated by the system.\n\n uid\t\n UID is the unique in time and space value for this object. It is typically\n generated by the server on successful creation of a resource and is not\n allowed to change on PUT operations.\n\n Populated by the system. Read-only. More info:\n http://kubernetes.io/docs/user-guide/identifiers#uids\n\n" -Dec 14 16:20:12.232: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 explain e2e-test-crd-publish-openapi-3117-crds.spec' -Dec 14 16:20:12.706: INFO: stderr: "" -Dec 14 16:20:12.706: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-3117-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: spec \n\nDESCRIPTION:\n Specification of Foo\n\nFIELDS:\n bars\t<[]Object>\n List of Bars and their specs.\n\n" -Dec 14 16:20:12.708: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 explain e2e-test-crd-publish-openapi-3117-crds.spec.bars' -Dec 14 16:20:13.046: INFO: stderr: "" -Dec 14 16:20:13.046: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-3117-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: bars <[]Object>\n\nDESCRIPTION:\n List of Bars and their specs.\n\nFIELDS:\n age\t\n Age of Bar.\n\n bazs\t<[]string>\n List of Bazs.\n\n feeling\t\n Whether Bar is feeling great.\n\n name\t -required-\n Name of Bar.\n\n" -STEP: kubectl explain works to return error when explain is called on property that doesn't exist 12/14/22 16:20:13.046 -Dec 14 16:20:13.047: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 explain e2e-test-crd-publish-openapi-3117-crds.spec.bars2' -Dec 14 16:20:13.410: INFO: rc: 1 -[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - test/e2e/framework/framework.go:187 -Dec 14 16:20:16.463: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-publish-openapi-8852" for this suite. 12/14/22 16:20:16.476 -{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD with validation schema [Conformance]","completed":108,"skipped":1957,"failed":0} ------------------------------- -• [SLOW TEST] [13.105 seconds] -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - works for CRD with validation schema [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:68 +Feb 25 10:16:52.799: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Feb 25 10:16:52.799: INFO: update-demo-nautilus-8j946 is verified up and running +Feb 25 10:16:52.799: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-n9qzd -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' +Feb 25 10:16:52.997: INFO: stderr: "" +Feb 25 10:16:52.997: INFO: stdout: "true" +Feb 25 10:16:52.997: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-n9qzd -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' +Feb 25 10:16:53.167: INFO: stderr: "" +Feb 25 10:16:53.167: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" +Feb 25 10:16:53.167: INFO: validating pod update-demo-nautilus-n9qzd +Feb 25 10:16:53.209: INFO: got data: { + "image": "nautilus.jpg" +} - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:20:03.381 - Dec 14 16:20:03.381: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 16:20:03.383 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:20:03.415 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:20:03.419 - [It] works for CRD with validation schema [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:68 - Dec 14 16:20:03.424: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: kubectl validation (kubectl create and apply) allows request with known and required properties 12/14/22 16:20:06.524 - Dec 14 16:20:06.526: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 create -f -' - Dec 14 16:20:07.988: INFO: stderr: "" - Dec 14 16:20:07.988: INFO: stdout: "e2e-test-crd-publish-openapi-3117-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" - Dec 14 16:20:07.988: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 delete e2e-test-crd-publish-openapi-3117-crds test-foo' - Dec 14 16:20:08.272: INFO: stderr: "" - Dec 14 16:20:08.272: INFO: stdout: "e2e-test-crd-publish-openapi-3117-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" - Dec 14 16:20:08.273: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 apply -f -' - Dec 14 16:20:08.775: INFO: stderr: "" - Dec 14 16:20:08.775: INFO: stdout: "e2e-test-crd-publish-openapi-3117-crd.crd-publish-openapi-test-foo.example.com/test-foo created\n" - Dec 14 16:20:08.775: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 delete e2e-test-crd-publish-openapi-3117-crds test-foo' - Dec 14 16:20:08.908: INFO: stderr: "" - Dec 14 16:20:08.908: INFO: stdout: "e2e-test-crd-publish-openapi-3117-crd.crd-publish-openapi-test-foo.example.com \"test-foo\" deleted\n" - STEP: kubectl validation (kubectl create and apply) rejects request with value outside defined enum values 12/14/22 16:20:08.908 - Dec 14 16:20:08.909: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 create -f -' - Dec 14 16:20:09.711: INFO: rc: 1 - STEP: kubectl validation (kubectl create and apply) rejects request with unknown properties when disallowed by the schema 12/14/22 16:20:09.711 - Dec 14 16:20:09.711: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 create -f -' - Dec 14 16:20:10.136: INFO: rc: 1 - Dec 14 16:20:10.137: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 apply -f -' - Dec 14 16:20:10.659: INFO: rc: 1 - STEP: kubectl validation (kubectl create and apply) rejects request without required properties 12/14/22 16:20:10.659 - Dec 14 16:20:10.660: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 create -f -' - Dec 14 16:20:11.008: INFO: rc: 1 - Dec 14 16:20:11.008: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 --namespace=crd-publish-openapi-8852 apply -f -' - Dec 14 16:20:11.323: INFO: rc: 1 - STEP: kubectl explain works to explain CR properties 12/14/22 16:20:11.323 - Dec 14 16:20:11.324: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 explain e2e-test-crd-publish-openapi-3117-crds' - Dec 14 16:20:11.673: INFO: stderr: "" - Dec 14 16:20:11.673: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-3117-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nDESCRIPTION:\n Foo CRD for Testing\n\nFIELDS:\n apiVersion\t\n APIVersion defines the versioned schema of this representation of an\n object. Servers should convert recognized schemas to the latest internal\n value, and may reject unrecognized values. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n kind\t\n Kind is a string value representing the REST resource this object\n represents. Servers may infer this from the endpoint the client submits\n requests to. Cannot be updated. In CamelCase. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n metadata\t\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n spec\t\n Specification of Foo\n\n status\t\n Status of Foo\n\n" - STEP: kubectl explain works to explain CR properties recursively 12/14/22 16:20:11.674 - Dec 14 16:20:11.675: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 explain e2e-test-crd-publish-openapi-3117-crds.metadata' - Dec 14 16:20:12.231: INFO: stderr: "" - Dec 14 16:20:12.231: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-3117-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: metadata \n\nDESCRIPTION:\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n ObjectMeta is metadata that all persisted resources must have, which\n includes all objects users must create.\n\nFIELDS:\n annotations\t\n Annotations is an unstructured key value map stored with a resource that\n may be set by external tools to store and retrieve arbitrary metadata. They\n are not queryable and should be preserved when modifying objects. More\n info: http://kubernetes.io/docs/user-guide/annotations\n\n creationTimestamp\t\n CreationTimestamp is a timestamp representing the server time when this\n object was created. It is not guaranteed to be set in happens-before order\n across separate operations. Clients may not set this value. It is\n represented in RFC3339 form and is in UTC.\n\n Populated by the system. Read-only. Null for lists. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n deletionGracePeriodSeconds\t\n Number of seconds allowed for this object to gracefully terminate before it\n will be removed from the system. Only set when deletionTimestamp is also\n set. May only be shortened. Read-only.\n\n deletionTimestamp\t\n DeletionTimestamp is RFC 3339 date and time at which this resource will be\n deleted. This field is set by the server when a graceful deletion is\n requested by the user, and is not directly settable by a client. The\n resource is expected to be deleted (no longer visible from resource lists,\n and not reachable by name) after the time in this field, once the\n finalizers list is empty. As long as the finalizers list contains items,\n deletion is blocked. Once the deletionTimestamp is set, this value may not\n be unset or be set further into the future, although it may be shortened or\n the resource may be deleted prior to this time. For example, a user may\n request that a pod is deleted in 30 seconds. The Kubelet will react by\n sending a graceful termination signal to the containers in the pod. After\n that 30 seconds, the Kubelet will send a hard termination signal (SIGKILL)\n to the container and after cleanup, remove the pod from the API. In the\n presence of network partitions, this object may still exist after this\n timestamp, until an administrator or automated process can determine the\n resource is fully terminated. If not set, graceful deletion of the object\n has not been requested.\n\n Populated by the system when a graceful deletion is requested. Read-only.\n More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n finalizers\t<[]string>\n Must be empty before the object is deleted from the registry. Each entry is\n an identifier for the responsible component that will remove the entry from\n the list. If the deletionTimestamp of the object is non-nil, entries in\n this list can only be removed. Finalizers may be processed and removed in\n any order. Order is NOT enforced because it introduces significant risk of\n stuck finalizers. finalizers is a shared field, any actor with permission\n can reorder it. If the finalizer list is processed in order, then this can\n lead to a situation in which the component responsible for the first\n finalizer in the list is waiting for a signal (field value, external\n system, or other) produced by a component responsible for a finalizer later\n in the list, resulting in a deadlock. Without enforced ordering finalizers\n are free to order amongst themselves and are not vulnerable to ordering\n changes in the list.\n\n generateName\t\n GenerateName is an optional prefix, used by the server, to generate a\n unique name ONLY IF the Name field has not been provided. If this field is\n used, the name returned to the client will be different than the name\n passed. This value will also be combined with a unique suffix. The provided\n value has the same validation rules as the Name field, and may be truncated\n by the length of the suffix required to make the value unique on the\n server.\n\n If this field is specified and the generated name exists, the server will\n return a 409.\n\n Applied only if Name is not specified. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency\n\n generation\t\n A sequence number representing a specific generation of the desired state.\n Populated by the system. Read-only.\n\n labels\t\n Map of string keys and values that can be used to organize and categorize\n (scope and select) objects. May match selectors of replication controllers\n and services. More info: http://kubernetes.io/docs/user-guide/labels\n\n managedFields\t<[]Object>\n ManagedFields maps workflow-id and version to the set of fields that are\n managed by that workflow. This is mostly for internal housekeeping, and\n users typically shouldn't need to set or understand this field. A workflow\n can be the user's name, a controller's name, or the name of a specific\n apply path like \"ci-cd\". The set of fields is always in the version that\n the workflow used when modifying the object.\n\n name\t\n Name must be unique within a namespace. Is required when creating\n resources, although some resources may allow a client to request the\n generation of an appropriate name automatically. Name is primarily intended\n for creation idempotence and configuration definition. Cannot be updated.\n More info: http://kubernetes.io/docs/user-guide/identifiers#names\n\n namespace\t\n Namespace defines the space within which each name must be unique. An empty\n namespace is equivalent to the \"default\" namespace, but \"default\" is the\n canonical representation. Not all objects are required to be scoped to a\n namespace - the value of this field for those objects will be empty.\n\n Must be a DNS_LABEL. Cannot be updated. More info:\n http://kubernetes.io/docs/user-guide/namespaces\n\n ownerReferences\t<[]Object>\n List of objects depended by this object. If ALL objects in the list have\n been deleted, this object will be garbage collected. If this object is\n managed by a controller, then an entry in this list will point to this\n controller, with the controller field set to true. There cannot be more\n than one managing controller.\n\n resourceVersion\t\n An opaque value that represents the internal version of this object that\n can be used by clients to determine when objects have changed. May be used\n for optimistic concurrency, change detection, and the watch operation on a\n resource or set of resources. Clients must treat these values as opaque and\n passed unmodified back to the server. They may only be valid for a\n particular resource or set of resources.\n\n Populated by the system. Read-only. Value must be treated as opaque by\n clients and . More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency\n\n selfLink\t\n Deprecated: selfLink is a legacy read-only field that is no longer\n populated by the system.\n\n uid\t\n UID is the unique in time and space value for this object. It is typically\n generated by the server on successful creation of a resource and is not\n allowed to change on PUT operations.\n\n Populated by the system. Read-only. More info:\n http://kubernetes.io/docs/user-guide/identifiers#uids\n\n" - Dec 14 16:20:12.232: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 explain e2e-test-crd-publish-openapi-3117-crds.spec' - Dec 14 16:20:12.706: INFO: stderr: "" - Dec 14 16:20:12.706: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-3117-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: spec \n\nDESCRIPTION:\n Specification of Foo\n\nFIELDS:\n bars\t<[]Object>\n List of Bars and their specs.\n\n" - Dec 14 16:20:12.708: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 explain e2e-test-crd-publish-openapi-3117-crds.spec.bars' - Dec 14 16:20:13.046: INFO: stderr: "" - Dec 14 16:20:13.046: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-3117-crd\nVERSION: crd-publish-openapi-test-foo.example.com/v1\n\nRESOURCE: bars <[]Object>\n\nDESCRIPTION:\n List of Bars and their specs.\n\nFIELDS:\n age\t\n Age of Bar.\n\n bazs\t<[]string>\n List of Bazs.\n\n feeling\t\n Whether Bar is feeling great.\n\n name\t -required-\n Name of Bar.\n\n" - STEP: kubectl explain works to return error when explain is called on property that doesn't exist 12/14/22 16:20:13.046 - Dec 14 16:20:13.047: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-8852 explain e2e-test-crd-publish-openapi-3117-crds.spec.bars2' - Dec 14 16:20:13.410: INFO: rc: 1 - [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - test/e2e/framework/framework.go:187 - Dec 14 16:20:16.463: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "crd-publish-openapi-8852" for this suite. 12/14/22 16:20:16.476 - << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-node] Container Runtime blackbox test when starting a container that exits - should run with the expected status [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:51 -[BeforeEach] [sig-node] Container Runtime - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:20:16.498 -Dec 14 16:20:16.498: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-runtime 12/14/22 16:20:16.502 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:20:16.525 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:20:16.529 -[It] should run with the expected status [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:51 -STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount' 12/14/22 16:20:16.541 -STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase' 12/14/22 16:20:31.69 -STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition 12/14/22 16:20:31.697 -STEP: Container 'terminate-cmd-rpa': should get the expected 'State' 12/14/22 16:20:31.705 -STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance] 12/14/22 16:20:31.705 -STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount' 12/14/22 16:20:31.747 -STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase' 12/14/22 16:20:34.788 -STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition 12/14/22 16:20:36.812 -STEP: Container 'terminate-cmd-rpof': should get the expected 'State' 12/14/22 16:20:36.826 -STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance] 12/14/22 16:20:36.827 -STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount' 12/14/22 16:20:36.857 -STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase' 12/14/22 16:20:37.869 -STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition 12/14/22 16:20:39.894 -STEP: Container 'terminate-cmd-rpn': should get the expected 'State' 12/14/22 16:20:39.909 -STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance] 12/14/22 16:20:39.909 -[AfterEach] [sig-node] Container Runtime +Feb 25 10:16:53.209: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Feb 25 10:16:53.209: INFO: update-demo-nautilus-n9qzd is verified up and running +STEP: using delete to clean up resources 02/25/23 10:16:53.21 +Feb 25 10:16:53.211: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 delete --grace-period=0 --force -f -' +Feb 25 10:16:53.407: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Feb 25 10:16:53.407: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" +Feb 25 10:16:53.409: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get rc,svc -l name=update-demo --no-headers' +Feb 25 10:16:53.789: INFO: stderr: "No resources found in kubectl-6523 namespace.\n" +Feb 25 10:16:53.789: INFO: stdout: "" +Feb 25 10:16:53.790: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods -l name=update-demo -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' +Feb 25 10:16:54.101: INFO: stderr: "" +Feb 25 10:16:54.101: INFO: stdout: "" +[AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 -Dec 14 16:20:39.948: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-runtime-7681" for this suite. 12/14/22 16:20:39.957 -{"msg":"PASSED [sig-node] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance]","completed":109,"skipped":2003,"failed":0} +Feb 25 10:16:54.109: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-6523" for this suite. 02/25/23 10:16:54.143 +{"msg":"PASSED [sig-cli] Kubectl client Update Demo should create and stop a replication controller [Conformance]","completed":108,"skipped":2048,"failed":0} ------------------------------ -• [SLOW TEST] [23.472 seconds] -[sig-node] Container Runtime -test/e2e/common/node/framework.go:23 - blackbox test - test/e2e/common/node/runtime.go:43 - when starting a container that exits - test/e2e/common/node/runtime.go:44 - should run with the expected status [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:51 +• [SLOW TEST] [25.111 seconds] +[sig-cli] Kubectl client +test/e2e/kubectl/framework.go:23 + Update Demo + test/e2e/kubectl/kubectl.go:322 + should create and stop a replication controller [Conformance] + test/e2e/kubectl/kubectl.go:337 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Container Runtime + [BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:20:16.498 - Dec 14 16:20:16.498: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-runtime 12/14/22 16:20:16.502 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:20:16.525 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:20:16.529 - [It] should run with the expected status [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:51 - STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount' 12/14/22 16:20:16.541 - STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase' 12/14/22 16:20:31.69 - STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition 12/14/22 16:20:31.697 - STEP: Container 'terminate-cmd-rpa': should get the expected 'State' 12/14/22 16:20:31.705 - STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance] 12/14/22 16:20:31.705 - STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount' 12/14/22 16:20:31.747 - STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase' 12/14/22 16:20:34.788 - STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition 12/14/22 16:20:36.812 - STEP: Container 'terminate-cmd-rpof': should get the expected 'State' 12/14/22 16:20:36.826 - STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance] 12/14/22 16:20:36.827 - STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount' 12/14/22 16:20:36.857 - STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase' 12/14/22 16:20:37.869 - STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition 12/14/22 16:20:39.894 - STEP: Container 'terminate-cmd-rpn': should get the expected 'State' 12/14/22 16:20:39.909 - STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance] 12/14/22 16:20:39.909 - [AfterEach] [sig-node] Container Runtime - test/e2e/framework/framework.go:187 - Dec 14 16:20:39.948: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-runtime-7681" for this suite. 12/14/22 16:20:39.957 - << End Captured GinkgoWriter Output ------------------------------- -SS ------------------------------- -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - should unconditionally reject operations on fail closed webhook [Conformance] - test/e2e/apimachinery/webhook.go:238 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:20:39.975 -Dec 14 16:20:39.975: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 16:20:39.986 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:20:40.012 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:20:40.017 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 16:20:40.043 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:20:41.443 -STEP: Deploying the webhook pod 12/14/22 16:20:41.459 -STEP: Wait for the deployment to be ready 12/14/22 16:20:41.474 -Dec 14 16:20:41.491: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set -STEP: Deploying the webhook service 12/14/22 16:20:43.514 -STEP: Verifying the service has paired with the endpoint 12/14/22 16:20:43.531 -Dec 14 16:20:44.532: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] should unconditionally reject operations on fail closed webhook [Conformance] - test/e2e/apimachinery/webhook.go:238 -STEP: Registering a webhook that server cannot talk to, with fail closed policy, via the AdmissionRegistration API 12/14/22 16:20:44.543 -STEP: create a namespace for the webhook 12/14/22 16:20:44.58 -STEP: create a configmap should be unconditionally rejected by the webhook 12/14/22 16:20:44.594 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/framework/framework.go:187 -Dec 14 16:20:44.653: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-6868" for this suite. 12/14/22 16:20:44.661 -STEP: Destroying namespace "webhook-6868-markers" for this suite. 12/14/22 16:20:44.669 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance]","completed":110,"skipped":2005,"failed":0} ------------------------------- -• [4.774 seconds] -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - should unconditionally reject operations on fail closed webhook [Conformance] - test/e2e/apimachinery/webhook.go:238 + STEP: Creating a kubernetes client 02/25/23 10:16:29.057 + Feb 25 10:16:29.057: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 10:16:29.065 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:16:29.223 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:16:29.231 + [BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 + [BeforeEach] Update Demo + test/e2e/kubectl/kubectl.go:324 + [It] should create and stop a replication controller [Conformance] + test/e2e/kubectl/kubectl.go:337 + STEP: creating a replication controller 02/25/23 10:16:29.24 + Feb 25 10:16:29.241: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 create -f -' + Feb 25 10:16:30.020: INFO: stderr: "" + Feb 25 10:16:30.020: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" + STEP: waiting for all containers in name=update-demo pods to come up. 02/25/23 10:16:30.02 + Feb 25 10:16:30.022: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' + Feb 25 10:16:30.269: INFO: stderr: "" + Feb 25 10:16:30.269: INFO: stdout: "update-demo-nautilus-8j946 update-demo-nautilus-n9qzd " + Feb 25 10:16:30.275: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-8j946 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' + Feb 25 10:16:30.632: INFO: stderr: "" + Feb 25 10:16:30.632: INFO: stdout: "" + Feb 25 10:16:30.632: INFO: update-demo-nautilus-8j946 is created but not running + Feb 25 10:16:35.643: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' + Feb 25 10:16:35.946: INFO: stderr: "" + Feb 25 10:16:35.946: INFO: stdout: "update-demo-nautilus-8j946 update-demo-nautilus-n9qzd " + Feb 25 10:16:35.946: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-8j946 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' + Feb 25 10:16:36.270: INFO: stderr: "" + Feb 25 10:16:36.270: INFO: stdout: "" + Feb 25 10:16:36.270: INFO: update-demo-nautilus-8j946 is created but not running + Feb 25 10:16:41.272: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' + Feb 25 10:16:41.528: INFO: stderr: "" + Feb 25 10:16:41.529: INFO: stdout: "update-demo-nautilus-8j946 update-demo-nautilus-n9qzd " + Feb 25 10:16:41.529: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-8j946 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' + Feb 25 10:16:41.806: INFO: stderr: "" + Feb 25 10:16:41.806: INFO: stdout: "" + Feb 25 10:16:41.806: INFO: update-demo-nautilus-8j946 is created but not running + Feb 25 10:16:46.808: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' + Feb 25 10:16:47.045: INFO: stderr: "" + Feb 25 10:16:47.045: INFO: stdout: "update-demo-nautilus-8j946 update-demo-nautilus-n9qzd " + Feb 25 10:16:47.047: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-8j946 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' + Feb 25 10:16:47.293: INFO: stderr: "" + Feb 25 10:16:47.293: INFO: stdout: "" + Feb 25 10:16:47.293: INFO: update-demo-nautilus-8j946 is created but not running + Feb 25 10:16:52.294: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' + Feb 25 10:16:52.484: INFO: stderr: "" + Feb 25 10:16:52.484: INFO: stdout: "update-demo-nautilus-8j946 update-demo-nautilus-n9qzd " + Feb 25 10:16:52.484: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-8j946 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' + Feb 25 10:16:52.628: INFO: stderr: "" + Feb 25 10:16:52.628: INFO: stdout: "true" + Feb 25 10:16:52.628: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-8j946 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' + Feb 25 10:16:52.780: INFO: stderr: "" + Feb 25 10:16:52.780: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" + Feb 25 10:16:52.780: INFO: validating pod update-demo-nautilus-8j946 + Feb 25 10:16:52.797: INFO: got data: { + "image": "nautilus.jpg" + } - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:20:39.975 - Dec 14 16:20:39.975: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 16:20:39.986 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:20:40.012 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:20:40.017 - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 16:20:40.043 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:20:41.443 - STEP: Deploying the webhook pod 12/14/22 16:20:41.459 - STEP: Wait for the deployment to be ready 12/14/22 16:20:41.474 - Dec 14 16:20:41.491: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set - STEP: Deploying the webhook service 12/14/22 16:20:43.514 - STEP: Verifying the service has paired with the endpoint 12/14/22 16:20:43.531 - Dec 14 16:20:44.532: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] should unconditionally reject operations on fail closed webhook [Conformance] - test/e2e/apimachinery/webhook.go:238 - STEP: Registering a webhook that server cannot talk to, with fail closed policy, via the AdmissionRegistration API 12/14/22 16:20:44.543 - STEP: create a namespace for the webhook 12/14/22 16:20:44.58 - STEP: create a configmap should be unconditionally rejected by the webhook 12/14/22 16:20:44.594 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + Feb 25 10:16:52.799: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . + Feb 25 10:16:52.799: INFO: update-demo-nautilus-8j946 is verified up and running + Feb 25 10:16:52.799: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-n9qzd -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' + Feb 25 10:16:52.997: INFO: stderr: "" + Feb 25 10:16:52.997: INFO: stdout: "true" + Feb 25 10:16:52.997: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods update-demo-nautilus-n9qzd -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' + Feb 25 10:16:53.167: INFO: stderr: "" + Feb 25 10:16:53.167: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" + Feb 25 10:16:53.167: INFO: validating pod update-demo-nautilus-n9qzd + Feb 25 10:16:53.209: INFO: got data: { + "image": "nautilus.jpg" + } + + Feb 25 10:16:53.209: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . + Feb 25 10:16:53.209: INFO: update-demo-nautilus-n9qzd is verified up and running + STEP: using delete to clean up resources 02/25/23 10:16:53.21 + Feb 25 10:16:53.211: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 delete --grace-period=0 --force -f -' + Feb 25 10:16:53.407: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" + Feb 25 10:16:53.407: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" + Feb 25 10:16:53.409: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get rc,svc -l name=update-demo --no-headers' + Feb 25 10:16:53.789: INFO: stderr: "No resources found in kubectl-6523 namespace.\n" + Feb 25 10:16:53.789: INFO: stdout: "" + Feb 25 10:16:53.790: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6523 get pods -l name=update-demo -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' + Feb 25 10:16:54.101: INFO: stderr: "" + Feb 25 10:16:54.101: INFO: stdout: "" + [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 16:20:44.653: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-6868" for this suite. 12/14/22 16:20:44.661 - STEP: Destroying namespace "webhook-6868-markers" for this suite. 12/14/22 16:20:44.669 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 + Feb 25 10:16:54.109: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-6523" for this suite. 02/25/23 10:16:54.143 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSS +SSSSSS ------------------------------ -[sig-node] Kubelet when scheduling an agnhost Pod with hostAliases - should write entries to /etc/hosts [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:148 -[BeforeEach] [sig-node] Kubelet +[sig-cli] Kubectl client Guestbook application + should create and stop a working application [Conformance] + test/e2e/kubectl/kubectl.go:392 +[BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:20:44.757 -Dec 14 16:20:44.757: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubelet-test 12/14/22 16:20:44.76 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:20:44.788 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:20:44.794 -[BeforeEach] [sig-node] Kubelet - test/e2e/common/node/kubelet.go:41 -[It] should write entries to /etc/hosts [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:148 -STEP: Waiting for pod completion 12/14/22 16:20:44.809 -Dec 14 16:20:44.809: INFO: Waiting up to 3m0s for pod "agnhost-host-aliases6e5bca6c-8f00-430c-b07c-54fbcdc7411c" in namespace "kubelet-test-2743" to be "completed" -Dec 14 16:20:44.815: INFO: Pod "agnhost-host-aliases6e5bca6c-8f00-430c-b07c-54fbcdc7411c": Phase="Pending", Reason="", readiness=false. Elapsed: 5.04364ms -Dec 14 16:20:46.825: INFO: Pod "agnhost-host-aliases6e5bca6c-8f00-430c-b07c-54fbcdc7411c": Phase="Running", Reason="", readiness=false. Elapsed: 2.015896613s -Dec 14 16:20:48.823: INFO: Pod "agnhost-host-aliases6e5bca6c-8f00-430c-b07c-54fbcdc7411c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013806595s -Dec 14 16:20:48.823: INFO: Pod "agnhost-host-aliases6e5bca6c-8f00-430c-b07c-54fbcdc7411c" satisfied condition "completed" -[AfterEach] [sig-node] Kubelet - test/e2e/framework/framework.go:187 -Dec 14 16:20:48.846: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubelet-test-2743" for this suite. 12/14/22 16:20:48.852 -{"msg":"PASSED [sig-node] Kubelet when scheduling an agnhost Pod with hostAliases should write entries to /etc/hosts [NodeConformance] [Conformance]","completed":111,"skipped":2014,"failed":0} ------------------------------- -• [4.104 seconds] -[sig-node] Kubelet -test/e2e/common/node/framework.go:23 - when scheduling an agnhost Pod with hostAliases - test/e2e/common/node/kubelet.go:140 - should write entries to /etc/hosts [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:148 +STEP: Creating a kubernetes client 02/25/23 10:16:54.173 +Feb 25 10:16:54.174: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 10:16:54.181 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:16:54.277 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:16:54.296 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 +[It] should create and stop a working application [Conformance] + test/e2e/kubectl/kubectl.go:392 +STEP: creating all guestbook components 02/25/23 10:16:54.329 +Feb 25 10:16:54.330: INFO: apiVersion: v1 +kind: Service +metadata: + name: agnhost-replica + labels: + app: agnhost + role: replica + tier: backend +spec: + ports: + - port: 6379 + selector: + app: agnhost + role: replica + tier: backend - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Kubelet - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:20:44.757 - Dec 14 16:20:44.757: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubelet-test 12/14/22 16:20:44.76 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:20:44.788 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:20:44.794 - [BeforeEach] [sig-node] Kubelet - test/e2e/common/node/kubelet.go:41 - [It] should write entries to /etc/hosts [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:148 - STEP: Waiting for pod completion 12/14/22 16:20:44.809 - Dec 14 16:20:44.809: INFO: Waiting up to 3m0s for pod "agnhost-host-aliases6e5bca6c-8f00-430c-b07c-54fbcdc7411c" in namespace "kubelet-test-2743" to be "completed" - Dec 14 16:20:44.815: INFO: Pod "agnhost-host-aliases6e5bca6c-8f00-430c-b07c-54fbcdc7411c": Phase="Pending", Reason="", readiness=false. Elapsed: 5.04364ms - Dec 14 16:20:46.825: INFO: Pod "agnhost-host-aliases6e5bca6c-8f00-430c-b07c-54fbcdc7411c": Phase="Running", Reason="", readiness=false. Elapsed: 2.015896613s - Dec 14 16:20:48.823: INFO: Pod "agnhost-host-aliases6e5bca6c-8f00-430c-b07c-54fbcdc7411c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013806595s - Dec 14 16:20:48.823: INFO: Pod "agnhost-host-aliases6e5bca6c-8f00-430c-b07c-54fbcdc7411c" satisfied condition "completed" - [AfterEach] [sig-node] Kubelet - test/e2e/framework/framework.go:187 - Dec 14 16:20:48.846: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubelet-test-2743" for this suite. 12/14/22 16:20:48.852 - << End Captured GinkgoWriter Output ------------------------------- -SSSS ------------------------------- -[sig-apps] ReplicaSet - should validate Replicaset Status endpoints [Conformance] - test/e2e/apps/replica_set.go:176 -[BeforeEach] [sig-apps] ReplicaSet - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:20:48.867 -Dec 14 16:20:48.867: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename replicaset 12/14/22 16:20:48.869 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:20:48.89 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:20:48.896 -[It] should validate Replicaset Status endpoints [Conformance] - test/e2e/apps/replica_set.go:176 -STEP: Create a Replicaset 12/14/22 16:20:48.907 -STEP: Verify that the required pods have come up. 12/14/22 16:20:48.916 -Dec 14 16:20:48.921: INFO: Pod name sample-pod: Found 0 pods out of 1 -Dec 14 16:20:53.928: INFO: Pod name sample-pod: Found 1 pods out of 1 -STEP: ensuring each pod is running 12/14/22 16:20:53.929 -STEP: Getting /status 12/14/22 16:20:53.929 -Dec 14 16:20:53.936: INFO: Replicaset test-rs has Conditions: [] -STEP: updating the Replicaset Status 12/14/22 16:20:53.936 -Dec 14 16:20:53.956: INFO: updatedStatus.Conditions: []v1.ReplicaSetCondition{v1.ReplicaSetCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} -STEP: watching for the ReplicaSet status to be updated 12/14/22 16:20:53.957 -Dec 14 16:20:53.963: INFO: Observed &ReplicaSet event: ADDED -Dec 14 16:20:53.964: INFO: Observed &ReplicaSet event: MODIFIED -Dec 14 16:20:53.965: INFO: Observed &ReplicaSet event: MODIFIED -Dec 14 16:20:53.965: INFO: Observed &ReplicaSet event: MODIFIED -Dec 14 16:20:53.965: INFO: Found replicaset test-rs in namespace replicaset-931 with labels: map[name:sample-pod pod:httpd] annotations: map[] & Conditions: [{StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] -Dec 14 16:20:53.966: INFO: Replicaset test-rs has an updated status -STEP: patching the Replicaset Status 12/14/22 16:20:53.966 -Dec 14 16:20:53.967: INFO: Patch payload: {"status":{"conditions":[{"type":"StatusPatched","status":"True"}]}} -Dec 14 16:20:53.980: INFO: Patched status conditions: []v1.ReplicaSetCondition{v1.ReplicaSetCondition{Type:"StatusPatched", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"", Message:""}} -STEP: watching for the Replicaset status to be patched 12/14/22 16:20:53.98 -Dec 14 16:20:53.982: INFO: Observed &ReplicaSet event: ADDED -Dec 14 16:20:53.982: INFO: Observed &ReplicaSet event: MODIFIED -Dec 14 16:20:53.982: INFO: Observed &ReplicaSet event: MODIFIED -Dec 14 16:20:53.982: INFO: Observed &ReplicaSet event: MODIFIED -Dec 14 16:20:53.982: INFO: Observed replicaset test-rs in namespace replicaset-931 with annotations: map[] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} -Dec 14 16:20:53.983: INFO: Observed &ReplicaSet event: MODIFIED -Dec 14 16:20:53.983: INFO: Found replicaset test-rs in namespace replicaset-931 with labels: map[name:sample-pod pod:httpd] annotations: map[] & Conditions: {StatusPatched True 0001-01-01 00:00:00 +0000 UTC } -Dec 14 16:20:53.983: INFO: Replicaset test-rs has a patched status -[AfterEach] [sig-apps] ReplicaSet - test/e2e/framework/framework.go:187 -Dec 14 16:20:53.983: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "replicaset-931" for this suite. 12/14/22 16:20:53.994 -{"msg":"PASSED [sig-apps] ReplicaSet should validate Replicaset Status endpoints [Conformance]","completed":112,"skipped":2018,"failed":0} ------------------------------- -• [SLOW TEST] [5.136 seconds] -[sig-apps] ReplicaSet -test/e2e/apps/framework.go:23 - should validate Replicaset Status endpoints [Conformance] - test/e2e/apps/replica_set.go:176 +Feb 25 10:16:54.333: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 create -f -' +Feb 25 10:16:55.060: INFO: stderr: "" +Feb 25 10:16:55.060: INFO: stdout: "service/agnhost-replica created\n" +Feb 25 10:16:55.060: INFO: apiVersion: v1 +kind: Service +metadata: + name: agnhost-primary + labels: + app: agnhost + role: primary + tier: backend +spec: + ports: + - port: 6379 + targetPort: 6379 + selector: + app: agnhost + role: primary + tier: backend - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] ReplicaSet - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:20:48.867 - Dec 14 16:20:48.867: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename replicaset 12/14/22 16:20:48.869 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:20:48.89 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:20:48.896 - [It] should validate Replicaset Status endpoints [Conformance] - test/e2e/apps/replica_set.go:176 - STEP: Create a Replicaset 12/14/22 16:20:48.907 - STEP: Verify that the required pods have come up. 12/14/22 16:20:48.916 - Dec 14 16:20:48.921: INFO: Pod name sample-pod: Found 0 pods out of 1 - Dec 14 16:20:53.928: INFO: Pod name sample-pod: Found 1 pods out of 1 - STEP: ensuring each pod is running 12/14/22 16:20:53.929 - STEP: Getting /status 12/14/22 16:20:53.929 - Dec 14 16:20:53.936: INFO: Replicaset test-rs has Conditions: [] - STEP: updating the Replicaset Status 12/14/22 16:20:53.936 - Dec 14 16:20:53.956: INFO: updatedStatus.Conditions: []v1.ReplicaSetCondition{v1.ReplicaSetCondition{Type:"StatusUpdate", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} - STEP: watching for the ReplicaSet status to be updated 12/14/22 16:20:53.957 - Dec 14 16:20:53.963: INFO: Observed &ReplicaSet event: ADDED - Dec 14 16:20:53.964: INFO: Observed &ReplicaSet event: MODIFIED - Dec 14 16:20:53.965: INFO: Observed &ReplicaSet event: MODIFIED - Dec 14 16:20:53.965: INFO: Observed &ReplicaSet event: MODIFIED - Dec 14 16:20:53.965: INFO: Found replicaset test-rs in namespace replicaset-931 with labels: map[name:sample-pod pod:httpd] annotations: map[] & Conditions: [{StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] - Dec 14 16:20:53.966: INFO: Replicaset test-rs has an updated status - STEP: patching the Replicaset Status 12/14/22 16:20:53.966 - Dec 14 16:20:53.967: INFO: Patch payload: {"status":{"conditions":[{"type":"StatusPatched","status":"True"}]}} - Dec 14 16:20:53.980: INFO: Patched status conditions: []v1.ReplicaSetCondition{v1.ReplicaSetCondition{Type:"StatusPatched", Status:"True", LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"", Message:""}} - STEP: watching for the Replicaset status to be patched 12/14/22 16:20:53.98 - Dec 14 16:20:53.982: INFO: Observed &ReplicaSet event: ADDED - Dec 14 16:20:53.982: INFO: Observed &ReplicaSet event: MODIFIED - Dec 14 16:20:53.982: INFO: Observed &ReplicaSet event: MODIFIED - Dec 14 16:20:53.982: INFO: Observed &ReplicaSet event: MODIFIED - Dec 14 16:20:53.982: INFO: Observed replicaset test-rs in namespace replicaset-931 with annotations: map[] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} - Dec 14 16:20:53.983: INFO: Observed &ReplicaSet event: MODIFIED - Dec 14 16:20:53.983: INFO: Found replicaset test-rs in namespace replicaset-931 with labels: map[name:sample-pod pod:httpd] annotations: map[] & Conditions: {StatusPatched True 0001-01-01 00:00:00 +0000 UTC } - Dec 14 16:20:53.983: INFO: Replicaset test-rs has a patched status - [AfterEach] [sig-apps] ReplicaSet - test/e2e/framework/framework.go:187 - Dec 14 16:20:53.983: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "replicaset-931" for this suite. 12/14/22 16:20:53.994 - << End Captured GinkgoWriter Output ------------------------------- -SS ------------------------------- -[sig-node] PreStop - should call prestop when killing a pod [Conformance] - test/e2e/node/pre_stop.go:168 -[BeforeEach] [sig-node] PreStop - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:20:54.004 -Dec 14 16:20:54.004: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename prestop 12/14/22 16:20:54.007 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:20:54.031 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:20:54.035 -[BeforeEach] [sig-node] PreStop - test/e2e/node/pre_stop.go:159 -[It] should call prestop when killing a pod [Conformance] - test/e2e/node/pre_stop.go:168 -STEP: Creating server pod server in namespace prestop-3697 12/14/22 16:20:54.038 -STEP: Waiting for pods to come up. 12/14/22 16:20:54.052 -Dec 14 16:20:54.052: INFO: Waiting up to 5m0s for pod "server" in namespace "prestop-3697" to be "running" -Dec 14 16:20:54.058: INFO: Pod "server": Phase="Pending", Reason="", readiness=false. Elapsed: 6.064108ms -Dec 14 16:20:56.066: INFO: Pod "server": Phase="Running", Reason="", readiness=true. Elapsed: 2.01346593s -Dec 14 16:20:56.066: INFO: Pod "server" satisfied condition "running" -STEP: Creating tester pod tester in namespace prestop-3697 12/14/22 16:20:56.069 -Dec 14 16:20:56.078: INFO: Waiting up to 5m0s for pod "tester" in namespace "prestop-3697" to be "running" -Dec 14 16:20:56.083: INFO: Pod "tester": Phase="Pending", Reason="", readiness=false. Elapsed: 5.380393ms -Dec 14 16:20:58.090: INFO: Pod "tester": Phase="Running", Reason="", readiness=true. Elapsed: 2.012471828s -Dec 14 16:20:58.091: INFO: Pod "tester" satisfied condition "running" -STEP: Deleting pre-stop pod 12/14/22 16:20:58.091 -Dec 14 16:21:03.117: INFO: Saw: { - "Hostname": "server", - "Sent": null, - "Received": { - "prestop": 1 - }, - "Errors": null, - "Log": [ - "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", - "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up." - ], - "StillContactingPeers": true -} -STEP: Deleting the server pod 12/14/22 16:21:03.118 -[AfterEach] [sig-node] PreStop - test/e2e/framework/framework.go:187 -Dec 14 16:21:03.148: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "prestop-3697" for this suite. 12/14/22 16:21:03.163 -{"msg":"PASSED [sig-node] PreStop should call prestop when killing a pod [Conformance]","completed":113,"skipped":2020,"failed":0} ------------------------------- -• [SLOW TEST] [9.181 seconds] -[sig-node] PreStop -test/e2e/node/framework.go:23 - should call prestop when killing a pod [Conformance] - test/e2e/node/pre_stop.go:168 +Feb 25 10:16:55.060: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 create -f -' +Feb 25 10:16:55.829: INFO: stderr: "" +Feb 25 10:16:55.829: INFO: stdout: "service/agnhost-primary created\n" +Feb 25 10:16:55.832: INFO: apiVersion: v1 +kind: Service +metadata: + name: frontend + labels: + app: guestbook + tier: frontend +spec: + # if your cluster supports it, uncomment the following to automatically create + # an external load-balanced IP for the frontend service. + # type: LoadBalancer + ports: + - port: 80 + selector: + app: guestbook + tier: frontend - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] PreStop - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:20:54.004 - Dec 14 16:20:54.004: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename prestop 12/14/22 16:20:54.007 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:20:54.031 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:20:54.035 - [BeforeEach] [sig-node] PreStop - test/e2e/node/pre_stop.go:159 - [It] should call prestop when killing a pod [Conformance] - test/e2e/node/pre_stop.go:168 - STEP: Creating server pod server in namespace prestop-3697 12/14/22 16:20:54.038 - STEP: Waiting for pods to come up. 12/14/22 16:20:54.052 - Dec 14 16:20:54.052: INFO: Waiting up to 5m0s for pod "server" in namespace "prestop-3697" to be "running" - Dec 14 16:20:54.058: INFO: Pod "server": Phase="Pending", Reason="", readiness=false. Elapsed: 6.064108ms - Dec 14 16:20:56.066: INFO: Pod "server": Phase="Running", Reason="", readiness=true. Elapsed: 2.01346593s - Dec 14 16:20:56.066: INFO: Pod "server" satisfied condition "running" - STEP: Creating tester pod tester in namespace prestop-3697 12/14/22 16:20:56.069 - Dec 14 16:20:56.078: INFO: Waiting up to 5m0s for pod "tester" in namespace "prestop-3697" to be "running" - Dec 14 16:20:56.083: INFO: Pod "tester": Phase="Pending", Reason="", readiness=false. Elapsed: 5.380393ms - Dec 14 16:20:58.090: INFO: Pod "tester": Phase="Running", Reason="", readiness=true. Elapsed: 2.012471828s - Dec 14 16:20:58.091: INFO: Pod "tester" satisfied condition "running" - STEP: Deleting pre-stop pod 12/14/22 16:20:58.091 - Dec 14 16:21:03.117: INFO: Saw: { - "Hostname": "server", - "Sent": null, - "Received": { - "prestop": 1 - }, - "Errors": null, - "Log": [ - "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", - "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up." - ], - "StillContactingPeers": true - } - STEP: Deleting the server pod 12/14/22 16:21:03.118 - [AfterEach] [sig-node] PreStop - test/e2e/framework/framework.go:187 - Dec 14 16:21:03.148: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "prestop-3697" for this suite. 12/14/22 16:21:03.163 - << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSSSSSSSSS ------------------------------- -[sig-apps] Daemon set [Serial] - should update pod when spec was updated and update strategy is RollingUpdate [Conformance] - test/e2e/apps/daemon_set.go:373 -[BeforeEach] [sig-apps] Daemon set [Serial] - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:21:03.191 -Dec 14 16:21:03.191: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename daemonsets 12/14/22 16:21:03.193 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:03.212 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:03.215 -[BeforeEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:145 -[It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] - test/e2e/apps/daemon_set.go:373 -Dec 14 16:21:03.254: INFO: Creating simple daemon set daemon-set -STEP: Check that daemon pods launch on every node of the cluster. 12/14/22 16:21:03.266 -Dec 14 16:21:03.282: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:21:03.282: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 16:21:04.303: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:21:04.303: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 16:21:05.297: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 -Dec 14 16:21:05.297: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set -STEP: Update daemon pods image. 12/14/22 16:21:05.326 -STEP: Check that daemon pods images are updated. 12/14/22 16:21:05.357 -Dec 14 16:21:05.364: INFO: Wrong image for pod: daemon-set-27rvw. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. -Dec 14 16:21:05.364: INFO: Wrong image for pod: daemon-set-6dz6x. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. -Dec 14 16:21:05.364: INFO: Wrong image for pod: daemon-set-7mpnq. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. -Dec 14 16:21:06.394: INFO: Wrong image for pod: daemon-set-27rvw. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. -Dec 14 16:21:06.394: INFO: Wrong image for pod: daemon-set-6dz6x. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. -Dec 14 16:21:07.393: INFO: Wrong image for pod: daemon-set-27rvw. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. -Dec 14 16:21:07.393: INFO: Wrong image for pod: daemon-set-6dz6x. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. -Dec 14 16:21:08.394: INFO: Wrong image for pod: daemon-set-27rvw. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. -Dec 14 16:21:08.394: INFO: Wrong image for pod: daemon-set-6dz6x. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. -Dec 14 16:21:08.394: INFO: Pod daemon-set-855kk is not available -Dec 14 16:21:09.390: INFO: Wrong image for pod: daemon-set-27rvw. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. -Dec 14 16:21:10.394: INFO: Wrong image for pod: daemon-set-27rvw. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. -Dec 14 16:21:10.394: INFO: Pod daemon-set-l25vq is not available -Dec 14 16:21:12.391: INFO: Pod daemon-set-x8q4k is not available -STEP: Check that daemon pods are still running on every node of the cluster. 12/14/22 16:21:12.412 -Dec 14 16:21:12.432: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 -Dec 14 16:21:12.432: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 16:21:13.453: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 -Dec 14 16:21:13.453: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set -[AfterEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:110 -STEP: Deleting DaemonSet "daemon-set" 12/14/22 16:21:13.495 -STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6128, will wait for the garbage collector to delete the pods 12/14/22 16:21:13.496 -Dec 14 16:21:13.565: INFO: Deleting DaemonSet.extensions daemon-set took: 12.426342ms -Dec 14 16:21:13.666: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.649329ms -Dec 14 16:21:16.074: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:21:16.074: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set -Dec 14 16:21:16.078: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"15538"},"items":null} +Feb 25 10:16:55.835: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 create -f -' +Feb 25 10:16:56.833: INFO: stderr: "" +Feb 25 10:16:56.833: INFO: stdout: "service/frontend created\n" +Feb 25 10:16:56.833: INFO: apiVersion: apps/v1 +kind: Deployment +metadata: + name: frontend +spec: + replicas: 3 + selector: + matchLabels: + app: guestbook + tier: frontend + template: + metadata: + labels: + app: guestbook + tier: frontend + spec: + containers: + - name: guestbook-frontend + image: registry.k8s.io/e2e-test-images/agnhost:2.40 + args: [ "guestbook", "--backend-port", "6379" ] + resources: + requests: + cpu: 100m + memory: 100Mi + ports: + - containerPort: 80 + +Feb 25 10:16:56.834: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 create -f -' +Feb 25 10:16:57.515: INFO: stderr: "" +Feb 25 10:16:57.515: INFO: stdout: "deployment.apps/frontend created\n" +Feb 25 10:16:57.517: INFO: apiVersion: apps/v1 +kind: Deployment +metadata: + name: agnhost-primary +spec: + replicas: 1 + selector: + matchLabels: + app: agnhost + role: primary + tier: backend + template: + metadata: + labels: + app: agnhost + role: primary + tier: backend + spec: + containers: + - name: primary + image: registry.k8s.io/e2e-test-images/agnhost:2.40 + args: [ "guestbook", "--http-port", "6379" ] + resources: + requests: + cpu: 100m + memory: 100Mi + ports: + - containerPort: 6379 -Dec 14 16:21:16.084: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"15538"},"items":null} +Feb 25 10:16:57.518: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 create -f -' +Feb 25 10:16:58.730: INFO: stderr: "" +Feb 25 10:16:58.731: INFO: stdout: "deployment.apps/agnhost-primary created\n" +Feb 25 10:16:58.731: INFO: apiVersion: apps/v1 +kind: Deployment +metadata: + name: agnhost-replica +spec: + replicas: 2 + selector: + matchLabels: + app: agnhost + role: replica + tier: backend + template: + metadata: + labels: + app: agnhost + role: replica + tier: backend + spec: + containers: + - name: replica + image: registry.k8s.io/e2e-test-images/agnhost:2.40 + args: [ "guestbook", "--replicaof", "agnhost-primary", "--http-port", "6379" ] + resources: + requests: + cpu: 100m + memory: 100Mi + ports: + - containerPort: 6379 -[AfterEach] [sig-apps] Daemon set [Serial] +Feb 25 10:16:58.732: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 create -f -' +Feb 25 10:16:59.914: INFO: stderr: "" +Feb 25 10:16:59.914: INFO: stdout: "deployment.apps/agnhost-replica created\n" +STEP: validating guestbook app 02/25/23 10:16:59.914 +Feb 25 10:16:59.914: INFO: Waiting for all frontend pods to be Running. +Feb 25 10:17:04.965: INFO: Waiting for frontend to serve content. +Feb 25 10:17:05.007: INFO: Trying to add a new entry to the guestbook. +Feb 25 10:17:05.068: INFO: Verifying that added entry can be retrieved. +STEP: using delete to clean up resources 02/25/23 10:17:05.102 +Feb 25 10:17:05.103: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 delete --grace-period=0 --force -f -' +Feb 25 10:17:05.342: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Feb 25 10:17:05.342: INFO: stdout: "service \"agnhost-replica\" force deleted\n" +STEP: using delete to clean up resources 02/25/23 10:17:05.342 +Feb 25 10:17:05.345: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 delete --grace-period=0 --force -f -' +Feb 25 10:17:05.573: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Feb 25 10:17:05.573: INFO: stdout: "service \"agnhost-primary\" force deleted\n" +STEP: using delete to clean up resources 02/25/23 10:17:05.573 +Feb 25 10:17:05.575: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 delete --grace-period=0 --force -f -' +Feb 25 10:17:05.899: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Feb 25 10:17:05.899: INFO: stdout: "service \"frontend\" force deleted\n" +STEP: using delete to clean up resources 02/25/23 10:17:05.9 +Feb 25 10:17:05.900: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 delete --grace-period=0 --force -f -' +Feb 25 10:17:06.182: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Feb 25 10:17:06.182: INFO: stdout: "deployment.apps \"frontend\" force deleted\n" +STEP: using delete to clean up resources 02/25/23 10:17:06.182 +Feb 25 10:17:06.183: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 delete --grace-period=0 --force -f -' +Feb 25 10:17:06.488: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Feb 25 10:17:06.488: INFO: stdout: "deployment.apps \"agnhost-primary\" force deleted\n" +STEP: using delete to clean up resources 02/25/23 10:17:06.488 +Feb 25 10:17:06.494: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 delete --grace-period=0 --force -f -' +Feb 25 10:17:06.862: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Feb 25 10:17:06.862: INFO: stdout: "deployment.apps \"agnhost-replica\" force deleted\n" +[AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 -Dec 14 16:21:16.103: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "daemonsets-6128" for this suite. 12/14/22 16:21:16.107 -{"msg":"PASSED [sig-apps] Daemon set [Serial] should update pod when spec was updated and update strategy is RollingUpdate [Conformance]","completed":114,"skipped":2035,"failed":0} +Feb 25 10:17:06.864: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-7148" for this suite. 02/25/23 10:17:06.942 +{"msg":"PASSED [sig-cli] Kubectl client Guestbook application should create and stop a working application [Conformance]","completed":109,"skipped":2054,"failed":0} ------------------------------ -• [SLOW TEST] [12.924 seconds] -[sig-apps] Daemon set [Serial] -test/e2e/apps/framework.go:23 - should update pod when spec was updated and update strategy is RollingUpdate [Conformance] - test/e2e/apps/daemon_set.go:373 +• [SLOW TEST] [12.790 seconds] +[sig-cli] Kubectl client +test/e2e/kubectl/framework.go:23 + Guestbook application + test/e2e/kubectl/kubectl.go:367 + should create and stop a working application [Conformance] + test/e2e/kubectl/kubectl.go:392 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Daemon set [Serial] + [BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:21:03.191 - Dec 14 16:21:03.191: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename daemonsets 12/14/22 16:21:03.193 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:03.212 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:03.215 - [BeforeEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:145 - [It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] - test/e2e/apps/daemon_set.go:373 - Dec 14 16:21:03.254: INFO: Creating simple daemon set daemon-set - STEP: Check that daemon pods launch on every node of the cluster. 12/14/22 16:21:03.266 - Dec 14 16:21:03.282: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:21:03.282: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 16:21:04.303: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:21:04.303: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 16:21:05.297: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 - Dec 14 16:21:05.297: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set - STEP: Update daemon pods image. 12/14/22 16:21:05.326 - STEP: Check that daemon pods images are updated. 12/14/22 16:21:05.357 - Dec 14 16:21:05.364: INFO: Wrong image for pod: daemon-set-27rvw. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. - Dec 14 16:21:05.364: INFO: Wrong image for pod: daemon-set-6dz6x. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. - Dec 14 16:21:05.364: INFO: Wrong image for pod: daemon-set-7mpnq. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. - Dec 14 16:21:06.394: INFO: Wrong image for pod: daemon-set-27rvw. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. - Dec 14 16:21:06.394: INFO: Wrong image for pod: daemon-set-6dz6x. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. - Dec 14 16:21:07.393: INFO: Wrong image for pod: daemon-set-27rvw. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. - Dec 14 16:21:07.393: INFO: Wrong image for pod: daemon-set-6dz6x. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. - Dec 14 16:21:08.394: INFO: Wrong image for pod: daemon-set-27rvw. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. - Dec 14 16:21:08.394: INFO: Wrong image for pod: daemon-set-6dz6x. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. - Dec 14 16:21:08.394: INFO: Pod daemon-set-855kk is not available - Dec 14 16:21:09.390: INFO: Wrong image for pod: daemon-set-27rvw. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. - Dec 14 16:21:10.394: INFO: Wrong image for pod: daemon-set-27rvw. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. - Dec 14 16:21:10.394: INFO: Pod daemon-set-l25vq is not available - Dec 14 16:21:12.391: INFO: Pod daemon-set-x8q4k is not available - STEP: Check that daemon pods are still running on every node of the cluster. 12/14/22 16:21:12.412 - Dec 14 16:21:12.432: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 - Dec 14 16:21:12.432: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 16:21:13.453: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 - Dec 14 16:21:13.453: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set - [AfterEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:110 - STEP: Deleting DaemonSet "daemon-set" 12/14/22 16:21:13.495 - STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6128, will wait for the garbage collector to delete the pods 12/14/22 16:21:13.496 - Dec 14 16:21:13.565: INFO: Deleting DaemonSet.extensions daemon-set took: 12.426342ms - Dec 14 16:21:13.666: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.649329ms - Dec 14 16:21:16.074: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:21:16.074: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set - Dec 14 16:21:16.078: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"15538"},"items":null} + STEP: Creating a kubernetes client 02/25/23 10:16:54.173 + Feb 25 10:16:54.174: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 10:16:54.181 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:16:54.277 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:16:54.296 + [BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 + [It] should create and stop a working application [Conformance] + test/e2e/kubectl/kubectl.go:392 + STEP: creating all guestbook components 02/25/23 10:16:54.329 + Feb 25 10:16:54.330: INFO: apiVersion: v1 + kind: Service + metadata: + name: agnhost-replica + labels: + app: agnhost + role: replica + tier: backend + spec: + ports: + - port: 6379 + selector: + app: agnhost + role: replica + tier: backend - Dec 14 16:21:16.084: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"15538"},"items":null} + Feb 25 10:16:54.333: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 create -f -' + Feb 25 10:16:55.060: INFO: stderr: "" + Feb 25 10:16:55.060: INFO: stdout: "service/agnhost-replica created\n" + Feb 25 10:16:55.060: INFO: apiVersion: v1 + kind: Service + metadata: + name: agnhost-primary + labels: + app: agnhost + role: primary + tier: backend + spec: + ports: + - port: 6379 + targetPort: 6379 + selector: + app: agnhost + role: primary + tier: backend - [AfterEach] [sig-apps] Daemon set [Serial] - test/e2e/framework/framework.go:187 - Dec 14 16:21:16.103: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "daemonsets-6128" for this suite. 12/14/22 16:21:16.107 - << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSSSSSSS ------------------------------- -[sig-scheduling] SchedulerPredicates [Serial] - validates resource limits of pods that are allowed to run [Conformance] - test/e2e/scheduling/predicates.go:326 -[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:21:16.12 -Dec 14 16:21:16.120: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename sched-pred 12/14/22 16:21:16.125 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:16.16 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:16.163 -[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:92 -Dec 14 16:21:16.166: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready -Dec 14 16:21:16.180: INFO: Waiting for terminating namespaces to be deleted... -Dec 14 16:21:16.184: INFO: -Logging pods the apiserver thinks is on node kiedai4iakae-1 before test -Dec 14 16:21:16.194: INFO: kube-flannel-ds-d6vqv from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.194: INFO: Container kube-flannel ready: true, restart count 0 -Dec 14 16:21:16.194: INFO: coredns-565d847f94-6mrkc from kube-system started at 2022-12-14 16:08:53 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.194: INFO: Container coredns ready: true, restart count 0 -Dec 14 16:21:16.194: INFO: kube-addon-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.195: INFO: Container kube-addon-manager ready: true, restart count 1 -Dec 14 16:21:16.195: INFO: kube-apiserver-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.195: INFO: Container kube-apiserver ready: true, restart count 1 -Dec 14 16:21:16.195: INFO: kube-controller-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:29:20 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.195: INFO: Container kube-controller-manager ready: true, restart count 2 -Dec 14 16:21:16.195: INFO: kube-proxy-fs5fm from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.195: INFO: Container kube-proxy ready: true, restart count 0 -Dec 14 16:21:16.195: INFO: kube-scheduler-kiedai4iakae-1 from kube-system started at 2022-12-14 15:25:59 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.195: INFO: Container kube-scheduler ready: true, restart count 2 -Dec 14 16:21:16.195: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-z7zks from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 16:21:16.195: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 16:21:16.195: INFO: Container systemd-logs ready: true, restart count 0 -Dec 14 16:21:16.195: INFO: -Logging pods the apiserver thinks is on node kiedai4iakae-2 before test -Dec 14 16:21:16.208: INFO: kube-flannel-ds-nvdqm from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.208: INFO: Container kube-flannel ready: true, restart count 0 -Dec 14 16:21:16.208: INFO: coredns-565d847f94-xhv9r from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.208: INFO: Container coredns ready: true, restart count 0 -Dec 14 16:21:16.208: INFO: kube-addon-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.208: INFO: Container kube-addon-manager ready: true, restart count 1 -Dec 14 16:21:16.208: INFO: kube-apiserver-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.208: INFO: Container kube-apiserver ready: true, restart count 1 -Dec 14 16:21:16.208: INFO: kube-controller-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.208: INFO: Container kube-controller-manager ready: true, restart count 1 -Dec 14 16:21:16.208: INFO: kube-proxy-dqjw5 from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.208: INFO: Container kube-proxy ready: true, restart count 0 -Dec 14 16:21:16.208: INFO: kube-scheduler-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.208: INFO: Container kube-scheduler ready: true, restart count 2 -Dec 14 16:21:16.208: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-cgz78 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 16:21:16.208: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 16:21:16.208: INFO: Container systemd-logs ready: true, restart count 0 -Dec 14 16:21:16.208: INFO: -Logging pods the apiserver thinks is on node kiedai4iakae-3 before test -Dec 14 16:21:16.217: INFO: kube-flannel-ds-4ws4h from kube-flannel started at 2022-12-14 16:08:54 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.217: INFO: Container kube-flannel ready: true, restart count 0 -Dec 14 16:21:16.218: INFO: kube-proxy-8kpmb from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.218: INFO: Container kube-proxy ready: true, restart count 0 -Dec 14 16:21:16.218: INFO: tester from prestop-3697 started at 2022-12-14 16:20:56 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.218: INFO: Container tester ready: true, restart count 0 -Dec 14 16:21:16.218: INFO: sonobuoy from sonobuoy started at 2022-12-14 15:35:59 +0000 UTC (1 container statuses recorded) -Dec 14 16:21:16.218: INFO: Container kube-sonobuoy ready: true, restart count 0 -Dec 14 16:21:16.218: INFO: sonobuoy-e2e-job-177ae5e9cdd34bac from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 16:21:16.218: INFO: Container e2e ready: true, restart count 0 -Dec 14 16:21:16.218: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 16:21:16.218: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-2fkm5 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 16:21:16.218: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 16:21:16.218: INFO: Container systemd-logs ready: true, restart count 0 -[It] validates resource limits of pods that are allowed to run [Conformance] - test/e2e/scheduling/predicates.go:326 -STEP: verifying the node has the label node kiedai4iakae-1 12/14/22 16:21:16.247 -STEP: verifying the node has the label node kiedai4iakae-2 12/14/22 16:21:16.281 -STEP: verifying the node has the label node kiedai4iakae-3 12/14/22 16:21:16.304 -Dec 14 16:21:16.342: INFO: Pod kube-flannel-ds-4ws4h requesting resource cpu=100m on Node kiedai4iakae-3 -Dec 14 16:21:16.343: INFO: Pod kube-flannel-ds-d6vqv requesting resource cpu=100m on Node kiedai4iakae-1 -Dec 14 16:21:16.343: INFO: Pod kube-flannel-ds-nvdqm requesting resource cpu=100m on Node kiedai4iakae-2 -Dec 14 16:21:16.343: INFO: Pod coredns-565d847f94-6mrkc requesting resource cpu=100m on Node kiedai4iakae-1 -Dec 14 16:21:16.343: INFO: Pod coredns-565d847f94-xhv9r requesting resource cpu=100m on Node kiedai4iakae-2 -Dec 14 16:21:16.343: INFO: Pod kube-addon-manager-kiedai4iakae-1 requesting resource cpu=5m on Node kiedai4iakae-1 -Dec 14 16:21:16.343: INFO: Pod kube-addon-manager-kiedai4iakae-2 requesting resource cpu=5m on Node kiedai4iakae-2 -Dec 14 16:21:16.343: INFO: Pod kube-apiserver-kiedai4iakae-1 requesting resource cpu=250m on Node kiedai4iakae-1 -Dec 14 16:21:16.343: INFO: Pod kube-apiserver-kiedai4iakae-2 requesting resource cpu=250m on Node kiedai4iakae-2 -Dec 14 16:21:16.343: INFO: Pod kube-controller-manager-kiedai4iakae-1 requesting resource cpu=200m on Node kiedai4iakae-1 -Dec 14 16:21:16.343: INFO: Pod kube-controller-manager-kiedai4iakae-2 requesting resource cpu=200m on Node kiedai4iakae-2 -Dec 14 16:21:16.343: INFO: Pod kube-proxy-8kpmb requesting resource cpu=0m on Node kiedai4iakae-3 -Dec 14 16:21:16.343: INFO: Pod kube-proxy-dqjw5 requesting resource cpu=0m on Node kiedai4iakae-2 -Dec 14 16:21:16.343: INFO: Pod kube-proxy-fs5fm requesting resource cpu=0m on Node kiedai4iakae-1 -Dec 14 16:21:16.343: INFO: Pod kube-scheduler-kiedai4iakae-1 requesting resource cpu=100m on Node kiedai4iakae-1 -Dec 14 16:21:16.343: INFO: Pod kube-scheduler-kiedai4iakae-2 requesting resource cpu=100m on Node kiedai4iakae-2 -Dec 14 16:21:16.343: INFO: Pod tester requesting resource cpu=0m on Node kiedai4iakae-3 -Dec 14 16:21:16.343: INFO: Pod sonobuoy requesting resource cpu=0m on Node kiedai4iakae-3 -Dec 14 16:21:16.343: INFO: Pod sonobuoy-e2e-job-177ae5e9cdd34bac requesting resource cpu=0m on Node kiedai4iakae-3 -Dec 14 16:21:16.343: INFO: Pod sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-2fkm5 requesting resource cpu=0m on Node kiedai4iakae-3 -Dec 14 16:21:16.343: INFO: Pod sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-cgz78 requesting resource cpu=0m on Node kiedai4iakae-2 -Dec 14 16:21:16.343: INFO: Pod sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-z7zks requesting resource cpu=0m on Node kiedai4iakae-1 -STEP: Starting Pods to consume most of the cluster CPU. 12/14/22 16:21:16.343 -Dec 14 16:21:16.343: INFO: Creating a pod which consumes cpu=591m on Node kiedai4iakae-1 -Dec 14 16:21:16.365: INFO: Creating a pod which consumes cpu=591m on Node kiedai4iakae-2 -Dec 14 16:21:16.387: INFO: Creating a pod which consumes cpu=1050m on Node kiedai4iakae-3 -Dec 14 16:21:16.397: INFO: Waiting up to 5m0s for pod "filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328" in namespace "sched-pred-9946" to be "running" -Dec 14 16:21:16.407: INFO: Pod "filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328": Phase="Pending", Reason="", readiness=false. Elapsed: 9.85875ms -Dec 14 16:21:18.416: INFO: Pod "filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328": Phase="Running", Reason="", readiness=true. Elapsed: 2.019544496s -Dec 14 16:21:18.417: INFO: Pod "filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328" satisfied condition "running" -Dec 14 16:21:18.417: INFO: Waiting up to 5m0s for pod "filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530" in namespace "sched-pred-9946" to be "running" -Dec 14 16:21:18.422: INFO: Pod "filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530": Phase="Running", Reason="", readiness=true. Elapsed: 5.349407ms -Dec 14 16:21:18.422: INFO: Pod "filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530" satisfied condition "running" -Dec 14 16:21:18.422: INFO: Waiting up to 5m0s for pod "filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80" in namespace "sched-pred-9946" to be "running" -Dec 14 16:21:18.428: INFO: Pod "filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80": Phase="Running", Reason="", readiness=true. Elapsed: 5.855345ms -Dec 14 16:21:18.428: INFO: Pod "filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80" satisfied condition "running" -STEP: Creating another pod that requires unavailable amount of CPU. 12/14/22 16:21:18.428 -STEP: Considering event: -Type = [Normal], Name = [filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80.1730b55dd630d303], Reason = [Scheduled], Message = [Successfully assigned sched-pred-9946/filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80 to kiedai4iakae-3] 12/14/22 16:21:18.435 -STEP: Considering event: -Type = [Normal], Name = [filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80.1730b55dec9340ab], Reason = [Pulled], Message = [Container image "registry.k8s.io/pause:3.8" already present on machine] 12/14/22 16:21:18.435 -STEP: Considering event: -Type = [Normal], Name = [filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80.1730b55df34d96c8], Reason = [Created], Message = [Created container filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80] 12/14/22 16:21:18.435 -STEP: Considering event: -Type = [Normal], Name = [filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80.1730b55df4932a84], Reason = [Started], Message = [Started container filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80] 12/14/22 16:21:18.435 -STEP: Considering event: -Type = [Normal], Name = [filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530.1730b55dd62b6762], Reason = [Scheduled], Message = [Successfully assigned sched-pred-9946/filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530 to kiedai4iakae-2] 12/14/22 16:21:18.435 -STEP: Considering event: -Type = [Normal], Name = [filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530.1730b55df2db83e4], Reason = [Pulled], Message = [Container image "registry.k8s.io/pause:3.8" already present on machine] 12/14/22 16:21:18.435 -STEP: Considering event: -Type = [Normal], Name = [filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530.1730b55dfa648807], Reason = [Created], Message = [Created container filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530] 12/14/22 16:21:18.436 -STEP: Considering event: -Type = [Normal], Name = [filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530.1730b55dfc7c70a3], Reason = [Started], Message = [Started container filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530] 12/14/22 16:21:18.436 -STEP: Considering event: -Type = [Normal], Name = [filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328.1730b55dd448fdc7], Reason = [Scheduled], Message = [Successfully assigned sched-pred-9946/filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328 to kiedai4iakae-1] 12/14/22 16:21:18.436 -STEP: Considering event: -Type = [Normal], Name = [filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328.1730b55df59c17a4], Reason = [Pulled], Message = [Container image "registry.k8s.io/pause:3.8" already present on machine] 12/14/22 16:21:18.436 -STEP: Considering event: -Type = [Normal], Name = [filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328.1730b55e06a89cae], Reason = [Created], Message = [Created container filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328] 12/14/22 16:21:18.436 -STEP: Considering event: -Type = [Normal], Name = [filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328.1730b55e0909b106], Reason = [Started], Message = [Started container filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328] 12/14/22 16:21:18.436 -STEP: Considering event: -Type = [Warning], Name = [additional-pod.1730b55e5061a147], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 Insufficient cpu. preemption: 0/3 nodes are available: 3 No preemption victims found for incoming pod.] 12/14/22 16:21:18.467 -STEP: removing the label node off the node kiedai4iakae-2 12/14/22 16:21:19.469 -STEP: verifying the node doesn't have the label node 12/14/22 16:21:19.487 -STEP: removing the label node off the node kiedai4iakae-3 12/14/22 16:21:19.495 -STEP: verifying the node doesn't have the label node 12/14/22 16:21:19.511 -STEP: removing the label node off the node kiedai4iakae-1 12/14/22 16:21:19.52 -STEP: verifying the node doesn't have the label node 12/14/22 16:21:19.542 -[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/framework/framework.go:187 -Dec 14 16:21:19.550: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "sched-pred-9946" for this suite. 12/14/22 16:21:19.577 -[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:83 -{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run [Conformance]","completed":115,"skipped":2048,"failed":0} ------------------------------- -• [3.482 seconds] -[sig-scheduling] SchedulerPredicates [Serial] -test/e2e/scheduling/framework.go:40 - validates resource limits of pods that are allowed to run [Conformance] - test/e2e/scheduling/predicates.go:326 + Feb 25 10:16:55.060: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 create -f -' + Feb 25 10:16:55.829: INFO: stderr: "" + Feb 25 10:16:55.829: INFO: stdout: "service/agnhost-primary created\n" + Feb 25 10:16:55.832: INFO: apiVersion: v1 + kind: Service + metadata: + name: frontend + labels: + app: guestbook + tier: frontend + spec: + # if your cluster supports it, uncomment the following to automatically create + # an external load-balanced IP for the frontend service. + # type: LoadBalancer + ports: + - port: 80 + selector: + app: guestbook + tier: frontend + + Feb 25 10:16:55.835: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 create -f -' + Feb 25 10:16:56.833: INFO: stderr: "" + Feb 25 10:16:56.833: INFO: stdout: "service/frontend created\n" + Feb 25 10:16:56.833: INFO: apiVersion: apps/v1 + kind: Deployment + metadata: + name: frontend + spec: + replicas: 3 + selector: + matchLabels: + app: guestbook + tier: frontend + template: + metadata: + labels: + app: guestbook + tier: frontend + spec: + containers: + - name: guestbook-frontend + image: registry.k8s.io/e2e-test-images/agnhost:2.40 + args: [ "guestbook", "--backend-port", "6379" ] + resources: + requests: + cpu: 100m + memory: 100Mi + ports: + - containerPort: 80 + + Feb 25 10:16:56.834: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 create -f -' + Feb 25 10:16:57.515: INFO: stderr: "" + Feb 25 10:16:57.515: INFO: stdout: "deployment.apps/frontend created\n" + Feb 25 10:16:57.517: INFO: apiVersion: apps/v1 + kind: Deployment + metadata: + name: agnhost-primary + spec: + replicas: 1 + selector: + matchLabels: + app: agnhost + role: primary + tier: backend + template: + metadata: + labels: + app: agnhost + role: primary + tier: backend + spec: + containers: + - name: primary + image: registry.k8s.io/e2e-test-images/agnhost:2.40 + args: [ "guestbook", "--http-port", "6379" ] + resources: + requests: + cpu: 100m + memory: 100Mi + ports: + - containerPort: 6379 - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:21:16.12 - Dec 14 16:21:16.120: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename sched-pred 12/14/22 16:21:16.125 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:16.16 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:16.163 - [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:92 - Dec 14 16:21:16.166: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready - Dec 14 16:21:16.180: INFO: Waiting for terminating namespaces to be deleted... - Dec 14 16:21:16.184: INFO: - Logging pods the apiserver thinks is on node kiedai4iakae-1 before test - Dec 14 16:21:16.194: INFO: kube-flannel-ds-d6vqv from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.194: INFO: Container kube-flannel ready: true, restart count 0 - Dec 14 16:21:16.194: INFO: coredns-565d847f94-6mrkc from kube-system started at 2022-12-14 16:08:53 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.194: INFO: Container coredns ready: true, restart count 0 - Dec 14 16:21:16.194: INFO: kube-addon-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.195: INFO: Container kube-addon-manager ready: true, restart count 1 - Dec 14 16:21:16.195: INFO: kube-apiserver-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.195: INFO: Container kube-apiserver ready: true, restart count 1 - Dec 14 16:21:16.195: INFO: kube-controller-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:29:20 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.195: INFO: Container kube-controller-manager ready: true, restart count 2 - Dec 14 16:21:16.195: INFO: kube-proxy-fs5fm from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.195: INFO: Container kube-proxy ready: true, restart count 0 - Dec 14 16:21:16.195: INFO: kube-scheduler-kiedai4iakae-1 from kube-system started at 2022-12-14 15:25:59 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.195: INFO: Container kube-scheduler ready: true, restart count 2 - Dec 14 16:21:16.195: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-z7zks from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 16:21:16.195: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 16:21:16.195: INFO: Container systemd-logs ready: true, restart count 0 - Dec 14 16:21:16.195: INFO: - Logging pods the apiserver thinks is on node kiedai4iakae-2 before test - Dec 14 16:21:16.208: INFO: kube-flannel-ds-nvdqm from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.208: INFO: Container kube-flannel ready: true, restart count 0 - Dec 14 16:21:16.208: INFO: coredns-565d847f94-xhv9r from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.208: INFO: Container coredns ready: true, restart count 0 - Dec 14 16:21:16.208: INFO: kube-addon-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.208: INFO: Container kube-addon-manager ready: true, restart count 1 - Dec 14 16:21:16.208: INFO: kube-apiserver-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.208: INFO: Container kube-apiserver ready: true, restart count 1 - Dec 14 16:21:16.208: INFO: kube-controller-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.208: INFO: Container kube-controller-manager ready: true, restart count 1 - Dec 14 16:21:16.208: INFO: kube-proxy-dqjw5 from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.208: INFO: Container kube-proxy ready: true, restart count 0 - Dec 14 16:21:16.208: INFO: kube-scheduler-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.208: INFO: Container kube-scheduler ready: true, restart count 2 - Dec 14 16:21:16.208: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-cgz78 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 16:21:16.208: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 16:21:16.208: INFO: Container systemd-logs ready: true, restart count 0 - Dec 14 16:21:16.208: INFO: - Logging pods the apiserver thinks is on node kiedai4iakae-3 before test - Dec 14 16:21:16.217: INFO: kube-flannel-ds-4ws4h from kube-flannel started at 2022-12-14 16:08:54 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.217: INFO: Container kube-flannel ready: true, restart count 0 - Dec 14 16:21:16.218: INFO: kube-proxy-8kpmb from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.218: INFO: Container kube-proxy ready: true, restart count 0 - Dec 14 16:21:16.218: INFO: tester from prestop-3697 started at 2022-12-14 16:20:56 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.218: INFO: Container tester ready: true, restart count 0 - Dec 14 16:21:16.218: INFO: sonobuoy from sonobuoy started at 2022-12-14 15:35:59 +0000 UTC (1 container statuses recorded) - Dec 14 16:21:16.218: INFO: Container kube-sonobuoy ready: true, restart count 0 - Dec 14 16:21:16.218: INFO: sonobuoy-e2e-job-177ae5e9cdd34bac from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 16:21:16.218: INFO: Container e2e ready: true, restart count 0 - Dec 14 16:21:16.218: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 16:21:16.218: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-2fkm5 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 16:21:16.218: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 16:21:16.218: INFO: Container systemd-logs ready: true, restart count 0 - [It] validates resource limits of pods that are allowed to run [Conformance] - test/e2e/scheduling/predicates.go:326 - STEP: verifying the node has the label node kiedai4iakae-1 12/14/22 16:21:16.247 - STEP: verifying the node has the label node kiedai4iakae-2 12/14/22 16:21:16.281 - STEP: verifying the node has the label node kiedai4iakae-3 12/14/22 16:21:16.304 - Dec 14 16:21:16.342: INFO: Pod kube-flannel-ds-4ws4h requesting resource cpu=100m on Node kiedai4iakae-3 - Dec 14 16:21:16.343: INFO: Pod kube-flannel-ds-d6vqv requesting resource cpu=100m on Node kiedai4iakae-1 - Dec 14 16:21:16.343: INFO: Pod kube-flannel-ds-nvdqm requesting resource cpu=100m on Node kiedai4iakae-2 - Dec 14 16:21:16.343: INFO: Pod coredns-565d847f94-6mrkc requesting resource cpu=100m on Node kiedai4iakae-1 - Dec 14 16:21:16.343: INFO: Pod coredns-565d847f94-xhv9r requesting resource cpu=100m on Node kiedai4iakae-2 - Dec 14 16:21:16.343: INFO: Pod kube-addon-manager-kiedai4iakae-1 requesting resource cpu=5m on Node kiedai4iakae-1 - Dec 14 16:21:16.343: INFO: Pod kube-addon-manager-kiedai4iakae-2 requesting resource cpu=5m on Node kiedai4iakae-2 - Dec 14 16:21:16.343: INFO: Pod kube-apiserver-kiedai4iakae-1 requesting resource cpu=250m on Node kiedai4iakae-1 - Dec 14 16:21:16.343: INFO: Pod kube-apiserver-kiedai4iakae-2 requesting resource cpu=250m on Node kiedai4iakae-2 - Dec 14 16:21:16.343: INFO: Pod kube-controller-manager-kiedai4iakae-1 requesting resource cpu=200m on Node kiedai4iakae-1 - Dec 14 16:21:16.343: INFO: Pod kube-controller-manager-kiedai4iakae-2 requesting resource cpu=200m on Node kiedai4iakae-2 - Dec 14 16:21:16.343: INFO: Pod kube-proxy-8kpmb requesting resource cpu=0m on Node kiedai4iakae-3 - Dec 14 16:21:16.343: INFO: Pod kube-proxy-dqjw5 requesting resource cpu=0m on Node kiedai4iakae-2 - Dec 14 16:21:16.343: INFO: Pod kube-proxy-fs5fm requesting resource cpu=0m on Node kiedai4iakae-1 - Dec 14 16:21:16.343: INFO: Pod kube-scheduler-kiedai4iakae-1 requesting resource cpu=100m on Node kiedai4iakae-1 - Dec 14 16:21:16.343: INFO: Pod kube-scheduler-kiedai4iakae-2 requesting resource cpu=100m on Node kiedai4iakae-2 - Dec 14 16:21:16.343: INFO: Pod tester requesting resource cpu=0m on Node kiedai4iakae-3 - Dec 14 16:21:16.343: INFO: Pod sonobuoy requesting resource cpu=0m on Node kiedai4iakae-3 - Dec 14 16:21:16.343: INFO: Pod sonobuoy-e2e-job-177ae5e9cdd34bac requesting resource cpu=0m on Node kiedai4iakae-3 - Dec 14 16:21:16.343: INFO: Pod sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-2fkm5 requesting resource cpu=0m on Node kiedai4iakae-3 - Dec 14 16:21:16.343: INFO: Pod sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-cgz78 requesting resource cpu=0m on Node kiedai4iakae-2 - Dec 14 16:21:16.343: INFO: Pod sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-z7zks requesting resource cpu=0m on Node kiedai4iakae-1 - STEP: Starting Pods to consume most of the cluster CPU. 12/14/22 16:21:16.343 - Dec 14 16:21:16.343: INFO: Creating a pod which consumes cpu=591m on Node kiedai4iakae-1 - Dec 14 16:21:16.365: INFO: Creating a pod which consumes cpu=591m on Node kiedai4iakae-2 - Dec 14 16:21:16.387: INFO: Creating a pod which consumes cpu=1050m on Node kiedai4iakae-3 - Dec 14 16:21:16.397: INFO: Waiting up to 5m0s for pod "filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328" in namespace "sched-pred-9946" to be "running" - Dec 14 16:21:16.407: INFO: Pod "filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328": Phase="Pending", Reason="", readiness=false. Elapsed: 9.85875ms - Dec 14 16:21:18.416: INFO: Pod "filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328": Phase="Running", Reason="", readiness=true. Elapsed: 2.019544496s - Dec 14 16:21:18.417: INFO: Pod "filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328" satisfied condition "running" - Dec 14 16:21:18.417: INFO: Waiting up to 5m0s for pod "filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530" in namespace "sched-pred-9946" to be "running" - Dec 14 16:21:18.422: INFO: Pod "filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530": Phase="Running", Reason="", readiness=true. Elapsed: 5.349407ms - Dec 14 16:21:18.422: INFO: Pod "filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530" satisfied condition "running" - Dec 14 16:21:18.422: INFO: Waiting up to 5m0s for pod "filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80" in namespace "sched-pred-9946" to be "running" - Dec 14 16:21:18.428: INFO: Pod "filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80": Phase="Running", Reason="", readiness=true. Elapsed: 5.855345ms - Dec 14 16:21:18.428: INFO: Pod "filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80" satisfied condition "running" - STEP: Creating another pod that requires unavailable amount of CPU. 12/14/22 16:21:18.428 - STEP: Considering event: - Type = [Normal], Name = [filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80.1730b55dd630d303], Reason = [Scheduled], Message = [Successfully assigned sched-pred-9946/filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80 to kiedai4iakae-3] 12/14/22 16:21:18.435 - STEP: Considering event: - Type = [Normal], Name = [filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80.1730b55dec9340ab], Reason = [Pulled], Message = [Container image "registry.k8s.io/pause:3.8" already present on machine] 12/14/22 16:21:18.435 - STEP: Considering event: - Type = [Normal], Name = [filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80.1730b55df34d96c8], Reason = [Created], Message = [Created container filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80] 12/14/22 16:21:18.435 - STEP: Considering event: - Type = [Normal], Name = [filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80.1730b55df4932a84], Reason = [Started], Message = [Started container filler-pod-89c65e46-aefb-41f5-bc21-38b4a0dfec80] 12/14/22 16:21:18.435 - STEP: Considering event: - Type = [Normal], Name = [filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530.1730b55dd62b6762], Reason = [Scheduled], Message = [Successfully assigned sched-pred-9946/filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530 to kiedai4iakae-2] 12/14/22 16:21:18.435 - STEP: Considering event: - Type = [Normal], Name = [filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530.1730b55df2db83e4], Reason = [Pulled], Message = [Container image "registry.k8s.io/pause:3.8" already present on machine] 12/14/22 16:21:18.435 - STEP: Considering event: - Type = [Normal], Name = [filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530.1730b55dfa648807], Reason = [Created], Message = [Created container filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530] 12/14/22 16:21:18.436 - STEP: Considering event: - Type = [Normal], Name = [filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530.1730b55dfc7c70a3], Reason = [Started], Message = [Started container filler-pod-e7054953-f4ce-4597-8f42-1be3a53cd530] 12/14/22 16:21:18.436 - STEP: Considering event: - Type = [Normal], Name = [filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328.1730b55dd448fdc7], Reason = [Scheduled], Message = [Successfully assigned sched-pred-9946/filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328 to kiedai4iakae-1] 12/14/22 16:21:18.436 - STEP: Considering event: - Type = [Normal], Name = [filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328.1730b55df59c17a4], Reason = [Pulled], Message = [Container image "registry.k8s.io/pause:3.8" already present on machine] 12/14/22 16:21:18.436 - STEP: Considering event: - Type = [Normal], Name = [filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328.1730b55e06a89cae], Reason = [Created], Message = [Created container filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328] 12/14/22 16:21:18.436 - STEP: Considering event: - Type = [Normal], Name = [filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328.1730b55e0909b106], Reason = [Started], Message = [Started container filler-pod-f5190eec-57b8-497c-bac1-a5baae8df328] 12/14/22 16:21:18.436 - STEP: Considering event: - Type = [Warning], Name = [additional-pod.1730b55e5061a147], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 Insufficient cpu. preemption: 0/3 nodes are available: 3 No preemption victims found for incoming pod.] 12/14/22 16:21:18.467 - STEP: removing the label node off the node kiedai4iakae-2 12/14/22 16:21:19.469 - STEP: verifying the node doesn't have the label node 12/14/22 16:21:19.487 - STEP: removing the label node off the node kiedai4iakae-3 12/14/22 16:21:19.495 - STEP: verifying the node doesn't have the label node 12/14/22 16:21:19.511 - STEP: removing the label node off the node kiedai4iakae-1 12/14/22 16:21:19.52 - STEP: verifying the node doesn't have the label node 12/14/22 16:21:19.542 - [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/framework/framework.go:187 - Dec 14 16:21:19.550: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "sched-pred-9946" for this suite. 12/14/22 16:21:19.577 - [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:83 - << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSSSS ------------------------------- -[sig-storage] Downward API volume - should update labels on modification [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:129 -[BeforeEach] [sig-storage] Downward API volume - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:21:19.603 -Dec 14 16:21:19.603: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 16:21:19.606 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:19.628 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:19.633 -[BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 -[It] should update labels on modification [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:129 -STEP: Creating the pod 12/14/22 16:21:19.639 -Dec 14 16:21:19.658: INFO: Waiting up to 5m0s for pod "labelsupdateb1d81c91-5b7b-4f2a-a7ba-35557cf44dd5" in namespace "downward-api-606" to be "running and ready" -Dec 14 16:21:19.675: INFO: Pod "labelsupdateb1d81c91-5b7b-4f2a-a7ba-35557cf44dd5": Phase="Pending", Reason="", readiness=false. Elapsed: 16.364278ms -Dec 14 16:21:19.676: INFO: The phase of Pod labelsupdateb1d81c91-5b7b-4f2a-a7ba-35557cf44dd5 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:21:21.683: INFO: Pod "labelsupdateb1d81c91-5b7b-4f2a-a7ba-35557cf44dd5": Phase="Running", Reason="", readiness=true. Elapsed: 2.024351829s -Dec 14 16:21:21.683: INFO: The phase of Pod labelsupdateb1d81c91-5b7b-4f2a-a7ba-35557cf44dd5 is Running (Ready = true) -Dec 14 16:21:21.683: INFO: Pod "labelsupdateb1d81c91-5b7b-4f2a-a7ba-35557cf44dd5" satisfied condition "running and ready" -Dec 14 16:21:22.233: INFO: Successfully updated pod "labelsupdateb1d81c91-5b7b-4f2a-a7ba-35557cf44dd5" -[AfterEach] [sig-storage] Downward API volume - test/e2e/framework/framework.go:187 -Dec 14 16:21:26.263: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-606" for this suite. 12/14/22 16:21:26.269 -{"msg":"PASSED [sig-storage] Downward API volume should update labels on modification [NodeConformance] [Conformance]","completed":116,"skipped":2058,"failed":0} ------------------------------- -• [SLOW TEST] [6.675 seconds] -[sig-storage] Downward API volume -test/e2e/common/storage/framework.go:23 - should update labels on modification [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:129 + Feb 25 10:16:57.518: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 create -f -' + Feb 25 10:16:58.730: INFO: stderr: "" + Feb 25 10:16:58.731: INFO: stdout: "deployment.apps/agnhost-primary created\n" + Feb 25 10:16:58.731: INFO: apiVersion: apps/v1 + kind: Deployment + metadata: + name: agnhost-replica + spec: + replicas: 2 + selector: + matchLabels: + app: agnhost + role: replica + tier: backend + template: + metadata: + labels: + app: agnhost + role: replica + tier: backend + spec: + containers: + - name: replica + image: registry.k8s.io/e2e-test-images/agnhost:2.40 + args: [ "guestbook", "--replicaof", "agnhost-primary", "--http-port", "6379" ] + resources: + requests: + cpu: 100m + memory: 100Mi + ports: + - containerPort: 6379 - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Downward API volume - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:21:19.603 - Dec 14 16:21:19.603: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 16:21:19.606 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:19.628 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:19.633 - [BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 - [It] should update labels on modification [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:129 - STEP: Creating the pod 12/14/22 16:21:19.639 - Dec 14 16:21:19.658: INFO: Waiting up to 5m0s for pod "labelsupdateb1d81c91-5b7b-4f2a-a7ba-35557cf44dd5" in namespace "downward-api-606" to be "running and ready" - Dec 14 16:21:19.675: INFO: Pod "labelsupdateb1d81c91-5b7b-4f2a-a7ba-35557cf44dd5": Phase="Pending", Reason="", readiness=false. Elapsed: 16.364278ms - Dec 14 16:21:19.676: INFO: The phase of Pod labelsupdateb1d81c91-5b7b-4f2a-a7ba-35557cf44dd5 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:21:21.683: INFO: Pod "labelsupdateb1d81c91-5b7b-4f2a-a7ba-35557cf44dd5": Phase="Running", Reason="", readiness=true. Elapsed: 2.024351829s - Dec 14 16:21:21.683: INFO: The phase of Pod labelsupdateb1d81c91-5b7b-4f2a-a7ba-35557cf44dd5 is Running (Ready = true) - Dec 14 16:21:21.683: INFO: Pod "labelsupdateb1d81c91-5b7b-4f2a-a7ba-35557cf44dd5" satisfied condition "running and ready" - Dec 14 16:21:22.233: INFO: Successfully updated pod "labelsupdateb1d81c91-5b7b-4f2a-a7ba-35557cf44dd5" - [AfterEach] [sig-storage] Downward API volume + Feb 25 10:16:58.732: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 create -f -' + Feb 25 10:16:59.914: INFO: stderr: "" + Feb 25 10:16:59.914: INFO: stdout: "deployment.apps/agnhost-replica created\n" + STEP: validating guestbook app 02/25/23 10:16:59.914 + Feb 25 10:16:59.914: INFO: Waiting for all frontend pods to be Running. + Feb 25 10:17:04.965: INFO: Waiting for frontend to serve content. + Feb 25 10:17:05.007: INFO: Trying to add a new entry to the guestbook. + Feb 25 10:17:05.068: INFO: Verifying that added entry can be retrieved. + STEP: using delete to clean up resources 02/25/23 10:17:05.102 + Feb 25 10:17:05.103: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 delete --grace-period=0 --force -f -' + Feb 25 10:17:05.342: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" + Feb 25 10:17:05.342: INFO: stdout: "service \"agnhost-replica\" force deleted\n" + STEP: using delete to clean up resources 02/25/23 10:17:05.342 + Feb 25 10:17:05.345: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 delete --grace-period=0 --force -f -' + Feb 25 10:17:05.573: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" + Feb 25 10:17:05.573: INFO: stdout: "service \"agnhost-primary\" force deleted\n" + STEP: using delete to clean up resources 02/25/23 10:17:05.573 + Feb 25 10:17:05.575: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 delete --grace-period=0 --force -f -' + Feb 25 10:17:05.899: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" + Feb 25 10:17:05.899: INFO: stdout: "service \"frontend\" force deleted\n" + STEP: using delete to clean up resources 02/25/23 10:17:05.9 + Feb 25 10:17:05.900: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 delete --grace-period=0 --force -f -' + Feb 25 10:17:06.182: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" + Feb 25 10:17:06.182: INFO: stdout: "deployment.apps \"frontend\" force deleted\n" + STEP: using delete to clean up resources 02/25/23 10:17:06.182 + Feb 25 10:17:06.183: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 delete --grace-period=0 --force -f -' + Feb 25 10:17:06.488: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" + Feb 25 10:17:06.488: INFO: stdout: "deployment.apps \"agnhost-primary\" force deleted\n" + STEP: using delete to clean up resources 02/25/23 10:17:06.488 + Feb 25 10:17:06.494: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7148 delete --grace-period=0 --force -f -' + Feb 25 10:17:06.862: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" + Feb 25 10:17:06.862: INFO: stdout: "deployment.apps \"agnhost-replica\" force deleted\n" + [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 16:21:26.263: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-606" for this suite. 12/14/22 16:21:26.269 + Feb 25 10:17:06.864: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-7148" for this suite. 02/25/23 10:17:06.942 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] RuntimeClass - should schedule a Pod requesting a RuntimeClass without PodOverhead [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:104 -[BeforeEach] [sig-node] RuntimeClass +[sig-apps] DisruptionController Listing PodDisruptionBudgets for all namespaces + should list and delete a collection of PodDisruptionBudgets [Conformance] + test/e2e/apps/disruption.go:86 +[BeforeEach] [sig-apps] DisruptionController test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:21:26.284 -Dec 14 16:21:26.284: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename runtimeclass 12/14/22 16:21:26.287 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:26.31 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:26.315 -[It] should schedule a Pod requesting a RuntimeClass without PodOverhead [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:104 -Dec 14 16:21:26.341: INFO: Waiting up to 1m20s for at least 1 pods in namespace runtimeclass-6927 to be scheduled -[AfterEach] [sig-node] RuntimeClass +STEP: Creating a kubernetes client 02/25/23 10:17:06.974 +Feb 25 10:17:06.974: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename disruption 02/25/23 10:17:06.986 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:17:07.026 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:17:07.03 +[BeforeEach] [sig-apps] DisruptionController + test/e2e/apps/disruption.go:71 +[BeforeEach] Listing PodDisruptionBudgets for all namespaces + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:17:07.041 +Feb 25 10:17:07.041: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename disruption-2 02/25/23 10:17:07.043 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:17:07.131 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:17:07.144 +[It] should list and delete a collection of PodDisruptionBudgets [Conformance] + test/e2e/apps/disruption.go:86 +STEP: Waiting for the pdb to be processed 02/25/23 10:17:07.181 +STEP: Waiting for the pdb to be processed 02/25/23 10:17:09.25 +STEP: Waiting for the pdb to be processed 02/25/23 10:17:11.278 +STEP: listing a collection of PDBs across all namespaces 02/25/23 10:17:13.3 +STEP: listing a collection of PDBs in namespace disruption-3670 02/25/23 10:17:13.308 +STEP: deleting a collection of PDBs 02/25/23 10:17:13.316 +STEP: Waiting for the PDB collection to be deleted 02/25/23 10:17:13.342 +[AfterEach] Listing PodDisruptionBudgets for all namespaces test/e2e/framework/framework.go:187 -Dec 14 16:21:26.365: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "runtimeclass-6927" for this suite. 12/14/22 16:21:26.378 -{"msg":"PASSED [sig-node] RuntimeClass should schedule a Pod requesting a RuntimeClass without PodOverhead [NodeConformance] [Conformance]","completed":117,"skipped":2083,"failed":0} +Feb 25 10:17:13.349: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "disruption-2-7999" for this suite. 02/25/23 10:17:13.358 +[AfterEach] [sig-apps] DisruptionController + test/e2e/framework/framework.go:187 +Feb 25 10:17:13.375: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "disruption-3670" for this suite. 02/25/23 10:17:13.385 +{"msg":"PASSED [sig-apps] DisruptionController Listing PodDisruptionBudgets for all namespaces should list and delete a collection of PodDisruptionBudgets [Conformance]","completed":110,"skipped":2089,"failed":0} ------------------------------ -• [0.105 seconds] -[sig-node] RuntimeClass -test/e2e/common/node/framework.go:23 - should schedule a Pod requesting a RuntimeClass without PodOverhead [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:104 +• [SLOW TEST] [6.427 seconds] +[sig-apps] DisruptionController +test/e2e/apps/framework.go:23 + Listing PodDisruptionBudgets for all namespaces + test/e2e/apps/disruption.go:77 + should list and delete a collection of PodDisruptionBudgets [Conformance] + test/e2e/apps/disruption.go:86 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] RuntimeClass + [BeforeEach] [sig-apps] DisruptionController test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:21:26.284 - Dec 14 16:21:26.284: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename runtimeclass 12/14/22 16:21:26.287 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:26.31 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:26.315 - [It] should schedule a Pod requesting a RuntimeClass without PodOverhead [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:104 - Dec 14 16:21:26.341: INFO: Waiting up to 1m20s for at least 1 pods in namespace runtimeclass-6927 to be scheduled - [AfterEach] [sig-node] RuntimeClass + STEP: Creating a kubernetes client 02/25/23 10:17:06.974 + Feb 25 10:17:06.974: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename disruption 02/25/23 10:17:06.986 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:17:07.026 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:17:07.03 + [BeforeEach] [sig-apps] DisruptionController + test/e2e/apps/disruption.go:71 + [BeforeEach] Listing PodDisruptionBudgets for all namespaces + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:17:07.041 + Feb 25 10:17:07.041: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename disruption-2 02/25/23 10:17:07.043 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:17:07.131 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:17:07.144 + [It] should list and delete a collection of PodDisruptionBudgets [Conformance] + test/e2e/apps/disruption.go:86 + STEP: Waiting for the pdb to be processed 02/25/23 10:17:07.181 + STEP: Waiting for the pdb to be processed 02/25/23 10:17:09.25 + STEP: Waiting for the pdb to be processed 02/25/23 10:17:11.278 + STEP: listing a collection of PDBs across all namespaces 02/25/23 10:17:13.3 + STEP: listing a collection of PDBs in namespace disruption-3670 02/25/23 10:17:13.308 + STEP: deleting a collection of PDBs 02/25/23 10:17:13.316 + STEP: Waiting for the PDB collection to be deleted 02/25/23 10:17:13.342 + [AfterEach] Listing PodDisruptionBudgets for all namespaces test/e2e/framework/framework.go:187 - Dec 14 16:21:26.365: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "runtimeclass-6927" for this suite. 12/14/22 16:21:26.378 + Feb 25 10:17:13.349: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "disruption-2-7999" for this suite. 02/25/23 10:17:13.358 + [AfterEach] [sig-apps] DisruptionController + test/e2e/framework/framework.go:187 + Feb 25 10:17:13.375: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "disruption-3670" for this suite. 02/25/23 10:17:13.385 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSS ------------------------------ -[sig-storage] EmptyDir wrapper volumes - should not conflict [Conformance] - test/e2e/storage/empty_dir_wrapper.go:67 -[BeforeEach] [sig-storage] EmptyDir wrapper volumes +[sig-apps] Deployment + Deployment should have a working scale subresource [Conformance] + test/e2e/apps/deployment.go:150 +[BeforeEach] [sig-apps] Deployment test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:21:26.398 -Dec 14 16:21:26.398: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir-wrapper 12/14/22 16:21:26.4 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:26.421 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:26.426 -[It] should not conflict [Conformance] - test/e2e/storage/empty_dir_wrapper.go:67 -Dec 14 16:21:26.461: INFO: Waiting up to 5m0s for pod "pod-secrets-abf4dbd7-4605-4e58-a944-dfa7c4620607" in namespace "emptydir-wrapper-6916" to be "running and ready" -Dec 14 16:21:26.483: INFO: Pod "pod-secrets-abf4dbd7-4605-4e58-a944-dfa7c4620607": Phase="Pending", Reason="", readiness=false. Elapsed: 21.649028ms -Dec 14 16:21:26.483: INFO: The phase of Pod pod-secrets-abf4dbd7-4605-4e58-a944-dfa7c4620607 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:21:28.493: INFO: Pod "pod-secrets-abf4dbd7-4605-4e58-a944-dfa7c4620607": Phase="Running", Reason="", readiness=true. Elapsed: 2.032024871s -Dec 14 16:21:28.494: INFO: The phase of Pod pod-secrets-abf4dbd7-4605-4e58-a944-dfa7c4620607 is Running (Ready = true) -Dec 14 16:21:28.494: INFO: Pod "pod-secrets-abf4dbd7-4605-4e58-a944-dfa7c4620607" satisfied condition "running and ready" -STEP: Cleaning up the secret 12/14/22 16:21:28.505 -STEP: Cleaning up the configmap 12/14/22 16:21:28.515 -STEP: Cleaning up the pod 12/14/22 16:21:28.523 -[AfterEach] [sig-storage] EmptyDir wrapper volumes +STEP: Creating a kubernetes client 02/25/23 10:17:13.409 +Feb 25 10:17:13.409: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename deployment 02/25/23 10:17:13.413 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:17:13.453 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:17:13.458 +[BeforeEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:91 +[It] Deployment should have a working scale subresource [Conformance] + test/e2e/apps/deployment.go:150 +Feb 25 10:17:13.462: INFO: Creating simple deployment test-new-deployment +Feb 25 10:17:13.493: INFO: deployment "test-new-deployment" doesn't have the required revision set +STEP: getting scale subresource 02/25/23 10:17:15.526 +STEP: updating a scale subresource 02/25/23 10:17:15.533 +STEP: verifying the deployment Spec.Replicas was modified 02/25/23 10:17:15.544 +STEP: Patch a scale subresource 02/25/23 10:17:15.553 +[AfterEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:84 +Feb 25 10:17:15.609: INFO: Deployment "test-new-deployment": +&Deployment{ObjectMeta:{test-new-deployment deployment-3423 2c28d84a-afa5-462b-b134-3662399b836e 17600 3 2023-02-25 10:17:13 +0000 UTC map[name:httpd] map[deployment.kubernetes.io/revision:1] [] [] [{e2e.test Update apps/v1 FieldsV1 {"f:spec":{"f:replicas":{}}} scale} {e2e.test Update apps/v1 2023-02-25 10:17:13 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:17:15 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*4,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc00246b408 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2023-02-25 10:17:15 +0000 UTC,LastTransitionTime:2023-02-25 10:17:15 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-new-deployment-845c8977d9" has successfully progressed.,LastUpdateTime:2023-02-25 10:17:15 +0000 UTC,LastTransitionTime:2023-02-25 10:17:13 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} + +Feb 25 10:17:15.618: INFO: New ReplicaSet "test-new-deployment-845c8977d9" of Deployment "test-new-deployment": +&ReplicaSet{ObjectMeta:{test-new-deployment-845c8977d9 deployment-3423 1e425f8d-a545-4b5b-970a-5da02b9410b3 17604 2 2023-02-25 10:17:13 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[deployment.kubernetes.io/desired-replicas:2 deployment.kubernetes.io/max-replicas:3 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-new-deployment 2c28d84a-afa5-462b-b134-3662399b836e 0xc004424d57 0xc004424d58}] [] [{kube-controller-manager Update apps/v1 2023-02-25 10:17:15 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"2c28d84a-afa5-462b-b134-3662399b836e\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:17:15 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*2,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 845c8977d9,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc004424de8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} +Feb 25 10:17:15.656: INFO: Pod "test-new-deployment-845c8977d9-2dwb4" is available: +&Pod{ObjectMeta:{test-new-deployment-845c8977d9-2dwb4 test-new-deployment-845c8977d9- deployment-3423 8182ad4e-a02e-4fe0-aafb-f03899854c2f 17594 0 2023-02-25 10:17:13 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet test-new-deployment-845c8977d9 1e425f8d-a545-4b5b-970a-5da02b9410b3 0xc00246b897 0xc00246b898}] [] [{kube-controller-manager Update v1 2023-02-25 10:17:13 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1e425f8d-a545-4b5b-970a-5da02b9410b3\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 10:17:15 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.17\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-fxtht,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fxtht,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:17:13 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:17:15 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:17:15 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:17:13 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:10.233.66.17,StartTime:2023-02-25 10:17:13 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 10:17:14 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://a4817530b4c6ad649e37e92e7a6779027b4cd1da8feceac88f67569f6b9551af,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.17,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 10:17:15.657: INFO: Pod "test-new-deployment-845c8977d9-7dnv9" is not available: +&Pod{ObjectMeta:{test-new-deployment-845c8977d9-7dnv9 test-new-deployment-845c8977d9- deployment-3423 caa22835-d19c-485a-b3d0-9fa05ec5214f 17605 0 2023-02-25 10:17:15 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet test-new-deployment-845c8977d9 1e425f8d-a545-4b5b-970a-5da02b9410b3 0xc00246ba87 0xc00246ba88}] [] [{kube-controller-manager Update v1 2023-02-25 10:17:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1e425f8d-a545-4b5b-970a-5da02b9410b3\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-9xc77,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9xc77,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:17:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} +[AfterEach] [sig-apps] Deployment test/e2e/framework/framework.go:187 -Dec 14 16:21:28.542: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-wrapper-6916" for this suite. 12/14/22 16:21:28.549 -{"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not conflict [Conformance]","completed":118,"skipped":2111,"failed":0} +Feb 25 10:17:15.657: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "deployment-3423" for this suite. 02/25/23 10:17:15.675 +{"msg":"PASSED [sig-apps] Deployment Deployment should have a working scale subresource [Conformance]","completed":111,"skipped":2098,"failed":0} ------------------------------ -• [2.160 seconds] -[sig-storage] EmptyDir wrapper volumes -test/e2e/storage/utils/framework.go:23 - should not conflict [Conformance] - test/e2e/storage/empty_dir_wrapper.go:67 +• [2.282 seconds] +[sig-apps] Deployment +test/e2e/apps/framework.go:23 + Deployment should have a working scale subresource [Conformance] + test/e2e/apps/deployment.go:150 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir wrapper volumes + [BeforeEach] [sig-apps] Deployment test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:21:26.398 - Dec 14 16:21:26.398: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir-wrapper 12/14/22 16:21:26.4 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:26.421 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:26.426 - [It] should not conflict [Conformance] - test/e2e/storage/empty_dir_wrapper.go:67 - Dec 14 16:21:26.461: INFO: Waiting up to 5m0s for pod "pod-secrets-abf4dbd7-4605-4e58-a944-dfa7c4620607" in namespace "emptydir-wrapper-6916" to be "running and ready" - Dec 14 16:21:26.483: INFO: Pod "pod-secrets-abf4dbd7-4605-4e58-a944-dfa7c4620607": Phase="Pending", Reason="", readiness=false. Elapsed: 21.649028ms - Dec 14 16:21:26.483: INFO: The phase of Pod pod-secrets-abf4dbd7-4605-4e58-a944-dfa7c4620607 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:21:28.493: INFO: Pod "pod-secrets-abf4dbd7-4605-4e58-a944-dfa7c4620607": Phase="Running", Reason="", readiness=true. Elapsed: 2.032024871s - Dec 14 16:21:28.494: INFO: The phase of Pod pod-secrets-abf4dbd7-4605-4e58-a944-dfa7c4620607 is Running (Ready = true) - Dec 14 16:21:28.494: INFO: Pod "pod-secrets-abf4dbd7-4605-4e58-a944-dfa7c4620607" satisfied condition "running and ready" - STEP: Cleaning up the secret 12/14/22 16:21:28.505 - STEP: Cleaning up the configmap 12/14/22 16:21:28.515 - STEP: Cleaning up the pod 12/14/22 16:21:28.523 - [AfterEach] [sig-storage] EmptyDir wrapper volumes + STEP: Creating a kubernetes client 02/25/23 10:17:13.409 + Feb 25 10:17:13.409: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename deployment 02/25/23 10:17:13.413 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:17:13.453 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:17:13.458 + [BeforeEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:91 + [It] Deployment should have a working scale subresource [Conformance] + test/e2e/apps/deployment.go:150 + Feb 25 10:17:13.462: INFO: Creating simple deployment test-new-deployment + Feb 25 10:17:13.493: INFO: deployment "test-new-deployment" doesn't have the required revision set + STEP: getting scale subresource 02/25/23 10:17:15.526 + STEP: updating a scale subresource 02/25/23 10:17:15.533 + STEP: verifying the deployment Spec.Replicas was modified 02/25/23 10:17:15.544 + STEP: Patch a scale subresource 02/25/23 10:17:15.553 + [AfterEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:84 + Feb 25 10:17:15.609: INFO: Deployment "test-new-deployment": + &Deployment{ObjectMeta:{test-new-deployment deployment-3423 2c28d84a-afa5-462b-b134-3662399b836e 17600 3 2023-02-25 10:17:13 +0000 UTC map[name:httpd] map[deployment.kubernetes.io/revision:1] [] [] [{e2e.test Update apps/v1 FieldsV1 {"f:spec":{"f:replicas":{}}} scale} {e2e.test Update apps/v1 2023-02-25 10:17:13 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:17:15 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*4,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc00246b408 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2023-02-25 10:17:15 +0000 UTC,LastTransitionTime:2023-02-25 10:17:15 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-new-deployment-845c8977d9" has successfully progressed.,LastUpdateTime:2023-02-25 10:17:15 +0000 UTC,LastTransitionTime:2023-02-25 10:17:13 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} + + Feb 25 10:17:15.618: INFO: New ReplicaSet "test-new-deployment-845c8977d9" of Deployment "test-new-deployment": + &ReplicaSet{ObjectMeta:{test-new-deployment-845c8977d9 deployment-3423 1e425f8d-a545-4b5b-970a-5da02b9410b3 17604 2 2023-02-25 10:17:13 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[deployment.kubernetes.io/desired-replicas:2 deployment.kubernetes.io/max-replicas:3 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-new-deployment 2c28d84a-afa5-462b-b134-3662399b836e 0xc004424d57 0xc004424d58}] [] [{kube-controller-manager Update apps/v1 2023-02-25 10:17:15 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"2c28d84a-afa5-462b-b134-3662399b836e\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:17:15 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*2,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 845c8977d9,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc004424de8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} + Feb 25 10:17:15.656: INFO: Pod "test-new-deployment-845c8977d9-2dwb4" is available: + &Pod{ObjectMeta:{test-new-deployment-845c8977d9-2dwb4 test-new-deployment-845c8977d9- deployment-3423 8182ad4e-a02e-4fe0-aafb-f03899854c2f 17594 0 2023-02-25 10:17:13 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet test-new-deployment-845c8977d9 1e425f8d-a545-4b5b-970a-5da02b9410b3 0xc00246b897 0xc00246b898}] [] [{kube-controller-manager Update v1 2023-02-25 10:17:13 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1e425f8d-a545-4b5b-970a-5da02b9410b3\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 10:17:15 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.17\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-fxtht,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fxtht,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:17:13 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:17:15 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:17:15 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:17:13 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:10.233.66.17,StartTime:2023-02-25 10:17:13 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 10:17:14 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://a4817530b4c6ad649e37e92e7a6779027b4cd1da8feceac88f67569f6b9551af,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.17,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 10:17:15.657: INFO: Pod "test-new-deployment-845c8977d9-7dnv9" is not available: + &Pod{ObjectMeta:{test-new-deployment-845c8977d9-7dnv9 test-new-deployment-845c8977d9- deployment-3423 caa22835-d19c-485a-b3d0-9fa05ec5214f 17605 0 2023-02-25 10:17:15 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet test-new-deployment-845c8977d9 1e425f8d-a545-4b5b-970a-5da02b9410b3 0xc00246ba87 0xc00246ba88}] [] [{kube-controller-manager Update v1 2023-02-25 10:17:15 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1e425f8d-a545-4b5b-970a-5da02b9410b3\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-9xc77,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9xc77,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:17:15 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} + [AfterEach] [sig-apps] Deployment test/e2e/framework/framework.go:187 - Dec 14 16:21:28.542: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-wrapper-6916" for this suite. 12/14/22 16:21:28.549 + Feb 25 10:17:15.657: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "deployment-3423" for this suite. 02/25/23 10:17:15.675 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSS +SS ------------------------------ -[sig-node] ConfigMap - should be consumable via the environment [NodeConformance] [Conformance] - test/e2e/common/node/configmap.go:92 -[BeforeEach] [sig-node] ConfigMap +[sig-storage] EmptyDir volumes + volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:86 +[BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:21:28.567 -Dec 14 16:21:28.567: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename configmap 12/14/22 16:21:28.569 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:28.593 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:28.597 -[It] should be consumable via the environment [NodeConformance] [Conformance] - test/e2e/common/node/configmap.go:92 -STEP: Creating configMap configmap-3574/configmap-test-9c67ba24-b3a5-4ce8-9eb3-67ca1fc6ee54 12/14/22 16:21:28.601 -STEP: Creating a pod to test consume configMaps 12/14/22 16:21:28.608 -Dec 14 16:21:28.624: INFO: Waiting up to 5m0s for pod "pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f" in namespace "configmap-3574" to be "Succeeded or Failed" -Dec 14 16:21:28.631: INFO: Pod "pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f": Phase="Pending", Reason="", readiness=false. Elapsed: 7.413339ms -Dec 14 16:21:30.641: INFO: Pod "pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017745492s -Dec 14 16:21:32.641: INFO: Pod "pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.017231416s -STEP: Saw pod success 12/14/22 16:21:32.641 -Dec 14 16:21:32.641: INFO: Pod "pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f" satisfied condition "Succeeded or Failed" -Dec 14 16:21:32.648: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f container env-test: -STEP: delete the pod 12/14/22 16:21:32.66 -Dec 14 16:21:32.673: INFO: Waiting for pod pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f to disappear -Dec 14 16:21:32.678: INFO: Pod pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f no longer exists -[AfterEach] [sig-node] ConfigMap +STEP: Creating a kubernetes client 02/25/23 10:17:15.692 +Feb 25 10:17:15.692: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir 02/25/23 10:17:15.695 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:17:15.821 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:17:15.827 +[It] volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:86 +STEP: Creating a pod to test emptydir volume type on tmpfs 02/25/23 10:17:15.835 +Feb 25 10:17:15.855: INFO: Waiting up to 5m0s for pod "pod-571aa678-7442-4068-ba73-29c490277088" in namespace "emptydir-2783" to be "Succeeded or Failed" +Feb 25 10:17:15.867: INFO: Pod "pod-571aa678-7442-4068-ba73-29c490277088": Phase="Pending", Reason="", readiness=false. Elapsed: 11.698499ms +Feb 25 10:17:17.874: INFO: Pod "pod-571aa678-7442-4068-ba73-29c490277088": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018563639s +Feb 25 10:17:19.874: INFO: Pod "pod-571aa678-7442-4068-ba73-29c490277088": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018501685s +STEP: Saw pod success 02/25/23 10:17:19.874 +Feb 25 10:17:19.874: INFO: Pod "pod-571aa678-7442-4068-ba73-29c490277088" satisfied condition "Succeeded or Failed" +Feb 25 10:17:19.880: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-571aa678-7442-4068-ba73-29c490277088 container test-container: +STEP: delete the pod 02/25/23 10:17:19.909 +Feb 25 10:17:19.928: INFO: Waiting for pod pod-571aa678-7442-4068-ba73-29c490277088 to disappear +Feb 25 10:17:19.934: INFO: Pod pod-571aa678-7442-4068-ba73-29c490277088 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 -Dec 14 16:21:32.679: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-3574" for this suite. 12/14/22 16:21:32.685 -{"msg":"PASSED [sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance]","completed":119,"skipped":2138,"failed":0} +Feb 25 10:17:19.935: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-2783" for this suite. 02/25/23 10:17:19.944 +{"msg":"PASSED [sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","completed":112,"skipped":2100,"failed":0} ------------------------------ -• [4.128 seconds] -[sig-node] ConfigMap -test/e2e/common/node/framework.go:23 - should be consumable via the environment [NodeConformance] [Conformance] - test/e2e/common/node/configmap.go:92 +• [4.264 seconds] +[sig-storage] EmptyDir volumes +test/e2e/common/storage/framework.go:23 + volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:86 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] ConfigMap + [BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:21:28.567 - Dec 14 16:21:28.567: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename configmap 12/14/22 16:21:28.569 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:28.593 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:28.597 - [It] should be consumable via the environment [NodeConformance] [Conformance] - test/e2e/common/node/configmap.go:92 - STEP: Creating configMap configmap-3574/configmap-test-9c67ba24-b3a5-4ce8-9eb3-67ca1fc6ee54 12/14/22 16:21:28.601 - STEP: Creating a pod to test consume configMaps 12/14/22 16:21:28.608 - Dec 14 16:21:28.624: INFO: Waiting up to 5m0s for pod "pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f" in namespace "configmap-3574" to be "Succeeded or Failed" - Dec 14 16:21:28.631: INFO: Pod "pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f": Phase="Pending", Reason="", readiness=false. Elapsed: 7.413339ms - Dec 14 16:21:30.641: INFO: Pod "pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017745492s - Dec 14 16:21:32.641: INFO: Pod "pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.017231416s - STEP: Saw pod success 12/14/22 16:21:32.641 - Dec 14 16:21:32.641: INFO: Pod "pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f" satisfied condition "Succeeded or Failed" - Dec 14 16:21:32.648: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f container env-test: - STEP: delete the pod 12/14/22 16:21:32.66 - Dec 14 16:21:32.673: INFO: Waiting for pod pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f to disappear - Dec 14 16:21:32.678: INFO: Pod pod-configmaps-f3e6cdeb-4c55-4655-8a66-9cdd7ee3267f no longer exists - [AfterEach] [sig-node] ConfigMap + STEP: Creating a kubernetes client 02/25/23 10:17:15.692 + Feb 25 10:17:15.692: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir 02/25/23 10:17:15.695 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:17:15.821 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:17:15.827 + [It] volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:86 + STEP: Creating a pod to test emptydir volume type on tmpfs 02/25/23 10:17:15.835 + Feb 25 10:17:15.855: INFO: Waiting up to 5m0s for pod "pod-571aa678-7442-4068-ba73-29c490277088" in namespace "emptydir-2783" to be "Succeeded or Failed" + Feb 25 10:17:15.867: INFO: Pod "pod-571aa678-7442-4068-ba73-29c490277088": Phase="Pending", Reason="", readiness=false. Elapsed: 11.698499ms + Feb 25 10:17:17.874: INFO: Pod "pod-571aa678-7442-4068-ba73-29c490277088": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018563639s + Feb 25 10:17:19.874: INFO: Pod "pod-571aa678-7442-4068-ba73-29c490277088": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018501685s + STEP: Saw pod success 02/25/23 10:17:19.874 + Feb 25 10:17:19.874: INFO: Pod "pod-571aa678-7442-4068-ba73-29c490277088" satisfied condition "Succeeded or Failed" + Feb 25 10:17:19.880: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-571aa678-7442-4068-ba73-29c490277088 container test-container: + STEP: delete the pod 02/25/23 10:17:19.909 + Feb 25 10:17:19.928: INFO: Waiting for pod pod-571aa678-7442-4068-ba73-29c490277088 to disappear + Feb 25 10:17:19.934: INFO: Pod pod-571aa678-7442-4068-ba73-29c490277088 no longer exists + [AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 - Dec 14 16:21:32.679: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "configmap-3574" for this suite. 12/14/22 16:21:32.685 + Feb 25 10:17:19.935: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-2783" for this suite. 02/25/23 10:17:19.944 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +S ------------------------------ -[sig-storage] Projected downwardAPI - should update labels on modification [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:129 -[BeforeEach] [sig-storage] Projected downwardAPI +[sig-scheduling] LimitRange + should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] + test/e2e/scheduling/limit_range.go:57 +[BeforeEach] [sig-scheduling] LimitRange test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:21:32.703 -Dec 14 16:21:32.703: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 16:21:32.704 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:32.725 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:32.729 -[BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 -[It] should update labels on modification [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:129 -STEP: Creating the pod 12/14/22 16:21:32.732 -Dec 14 16:21:32.745: INFO: Waiting up to 5m0s for pod "labelsupdatef0a13605-7897-4486-86f3-174b98556bed" in namespace "projected-5765" to be "running and ready" -Dec 14 16:21:32.749: INFO: Pod "labelsupdatef0a13605-7897-4486-86f3-174b98556bed": Phase="Pending", Reason="", readiness=false. Elapsed: 4.713893ms -Dec 14 16:21:32.750: INFO: The phase of Pod labelsupdatef0a13605-7897-4486-86f3-174b98556bed is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:21:34.762: INFO: Pod "labelsupdatef0a13605-7897-4486-86f3-174b98556bed": Phase="Running", Reason="", readiness=true. Elapsed: 2.017569104s -Dec 14 16:21:34.762: INFO: The phase of Pod labelsupdatef0a13605-7897-4486-86f3-174b98556bed is Running (Ready = true) -Dec 14 16:21:34.762: INFO: Pod "labelsupdatef0a13605-7897-4486-86f3-174b98556bed" satisfied condition "running and ready" -Dec 14 16:21:35.301: INFO: Successfully updated pod "labelsupdatef0a13605-7897-4486-86f3-174b98556bed" -[AfterEach] [sig-storage] Projected downwardAPI +STEP: Creating a kubernetes client 02/25/23 10:17:19.96 +Feb 25 10:17:19.960: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename limitrange 02/25/23 10:17:19.969 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:17:20.007 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:17:20.015 +[It] should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] + test/e2e/scheduling/limit_range.go:57 +STEP: Creating a LimitRange 02/25/23 10:17:20.019 +STEP: Setting up watch 02/25/23 10:17:20.02 +STEP: Submitting a LimitRange 02/25/23 10:17:20.127 +STEP: Verifying LimitRange creation was observed 02/25/23 10:17:20.137 +STEP: Fetching the LimitRange to ensure it has proper values 02/25/23 10:17:20.137 +Feb 25 10:17:20.143: INFO: Verifying requests: expected map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] with actual map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] +Feb 25 10:17:20.143: INFO: Verifying limits: expected map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] +STEP: Creating a Pod with no resource requirements 02/25/23 10:17:20.143 +STEP: Ensuring Pod has resource requirements applied from LimitRange 02/25/23 10:17:20.153 +Feb 25 10:17:20.161: INFO: Verifying requests: expected map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] with actual map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] +Feb 25 10:17:20.161: INFO: Verifying limits: expected map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] +STEP: Creating a Pod with partial resource requirements 02/25/23 10:17:20.161 +STEP: Ensuring Pod has merged resource requirements applied from LimitRange 02/25/23 10:17:20.175 +Feb 25 10:17:20.191: INFO: Verifying requests: expected map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{161061273600 0} {} 150Gi BinarySI} memory:{{157286400 0} {} 150Mi BinarySI}] with actual map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{161061273600 0} {} 150Gi BinarySI} memory:{{157286400 0} {} 150Mi BinarySI}] +Feb 25 10:17:20.191: INFO: Verifying limits: expected map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] +STEP: Failing to create a Pod with less than min resources 02/25/23 10:17:20.191 +STEP: Failing to create a Pod with more than max resources 02/25/23 10:17:20.196 +STEP: Updating a LimitRange 02/25/23 10:17:20.199 +STEP: Verifying LimitRange updating is effective 02/25/23 10:17:20.212 +STEP: Creating a Pod with less than former min resources 02/25/23 10:17:22.223 +STEP: Failing to create a Pod with more than max resources 02/25/23 10:17:22.233 +STEP: Deleting a LimitRange 02/25/23 10:17:22.237 +STEP: Verifying the LimitRange was deleted 02/25/23 10:17:22.25 +Feb 25 10:17:27.260: INFO: limitRange is already deleted +STEP: Creating a Pod with more than former max resources 02/25/23 10:17:27.26 +[AfterEach] [sig-scheduling] LimitRange test/e2e/framework/framework.go:187 -Dec 14 16:21:39.358: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-5765" for this suite. 12/14/22 16:21:39.368 -{"msg":"PASSED [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance]","completed":120,"skipped":2207,"failed":0} +Feb 25 10:17:27.282: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "limitrange-5763" for this suite. 02/25/23 10:17:27.296 +{"msg":"PASSED [sig-scheduling] LimitRange should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance]","completed":113,"skipped":2101,"failed":0} ------------------------------ -• [SLOW TEST] [6.680 seconds] -[sig-storage] Projected downwardAPI -test/e2e/common/storage/framework.go:23 - should update labels on modification [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:129 +• [SLOW TEST] [7.354 seconds] +[sig-scheduling] LimitRange +test/e2e/scheduling/framework.go:40 + should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] + test/e2e/scheduling/limit_range.go:57 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected downwardAPI + [BeforeEach] [sig-scheduling] LimitRange test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:21:32.703 - Dec 14 16:21:32.703: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 16:21:32.704 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:32.725 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:32.729 - [BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 - [It] should update labels on modification [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:129 - STEP: Creating the pod 12/14/22 16:21:32.732 - Dec 14 16:21:32.745: INFO: Waiting up to 5m0s for pod "labelsupdatef0a13605-7897-4486-86f3-174b98556bed" in namespace "projected-5765" to be "running and ready" - Dec 14 16:21:32.749: INFO: Pod "labelsupdatef0a13605-7897-4486-86f3-174b98556bed": Phase="Pending", Reason="", readiness=false. Elapsed: 4.713893ms - Dec 14 16:21:32.750: INFO: The phase of Pod labelsupdatef0a13605-7897-4486-86f3-174b98556bed is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:21:34.762: INFO: Pod "labelsupdatef0a13605-7897-4486-86f3-174b98556bed": Phase="Running", Reason="", readiness=true. Elapsed: 2.017569104s - Dec 14 16:21:34.762: INFO: The phase of Pod labelsupdatef0a13605-7897-4486-86f3-174b98556bed is Running (Ready = true) - Dec 14 16:21:34.762: INFO: Pod "labelsupdatef0a13605-7897-4486-86f3-174b98556bed" satisfied condition "running and ready" - Dec 14 16:21:35.301: INFO: Successfully updated pod "labelsupdatef0a13605-7897-4486-86f3-174b98556bed" - [AfterEach] [sig-storage] Projected downwardAPI + STEP: Creating a kubernetes client 02/25/23 10:17:19.96 + Feb 25 10:17:19.960: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename limitrange 02/25/23 10:17:19.969 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:17:20.007 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:17:20.015 + [It] should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] + test/e2e/scheduling/limit_range.go:57 + STEP: Creating a LimitRange 02/25/23 10:17:20.019 + STEP: Setting up watch 02/25/23 10:17:20.02 + STEP: Submitting a LimitRange 02/25/23 10:17:20.127 + STEP: Verifying LimitRange creation was observed 02/25/23 10:17:20.137 + STEP: Fetching the LimitRange to ensure it has proper values 02/25/23 10:17:20.137 + Feb 25 10:17:20.143: INFO: Verifying requests: expected map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] with actual map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] + Feb 25 10:17:20.143: INFO: Verifying limits: expected map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] + STEP: Creating a Pod with no resource requirements 02/25/23 10:17:20.143 + STEP: Ensuring Pod has resource requirements applied from LimitRange 02/25/23 10:17:20.153 + Feb 25 10:17:20.161: INFO: Verifying requests: expected map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] with actual map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] + Feb 25 10:17:20.161: INFO: Verifying limits: expected map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] + STEP: Creating a Pod with partial resource requirements 02/25/23 10:17:20.161 + STEP: Ensuring Pod has merged resource requirements applied from LimitRange 02/25/23 10:17:20.175 + Feb 25 10:17:20.191: INFO: Verifying requests: expected map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{161061273600 0} {} 150Gi BinarySI} memory:{{157286400 0} {} 150Mi BinarySI}] with actual map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{161061273600 0} {} 150Gi BinarySI} memory:{{157286400 0} {} 150Mi BinarySI}] + Feb 25 10:17:20.191: INFO: Verifying limits: expected map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] + STEP: Failing to create a Pod with less than min resources 02/25/23 10:17:20.191 + STEP: Failing to create a Pod with more than max resources 02/25/23 10:17:20.196 + STEP: Updating a LimitRange 02/25/23 10:17:20.199 + STEP: Verifying LimitRange updating is effective 02/25/23 10:17:20.212 + STEP: Creating a Pod with less than former min resources 02/25/23 10:17:22.223 + STEP: Failing to create a Pod with more than max resources 02/25/23 10:17:22.233 + STEP: Deleting a LimitRange 02/25/23 10:17:22.237 + STEP: Verifying the LimitRange was deleted 02/25/23 10:17:22.25 + Feb 25 10:17:27.260: INFO: limitRange is already deleted + STEP: Creating a Pod with more than former max resources 02/25/23 10:17:27.26 + [AfterEach] [sig-scheduling] LimitRange test/e2e/framework/framework.go:187 - Dec 14 16:21:39.358: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-5765" for this suite. 12/14/22 16:21:39.368 + Feb 25 10:17:27.282: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "limitrange-5763" for this suite. 02/25/23 10:17:27.296 << End Captured GinkgoWriter Output ------------------------------ -SS +SSSSSSSSSSSSS ------------------------------ -[sig-apps] Job - should adopt matching orphans and release non-matching pods [Conformance] - test/e2e/apps/job.go:335 -[BeforeEach] [sig-apps] Job +[sig-api-machinery] Garbage collector + should not be blocked by dependency circle [Conformance] + test/e2e/apimachinery/garbage_collector.go:849 +[BeforeEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:21:39.387 -Dec 14 16:21:39.388: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename job 12/14/22 16:21:39.392 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:39.412 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:39.417 -[It] should adopt matching orphans and release non-matching pods [Conformance] - test/e2e/apps/job.go:335 -STEP: Creating a job 12/14/22 16:21:39.42 -STEP: Ensuring active pods == parallelism 12/14/22 16:21:39.429 -STEP: Orphaning one of the Job's Pods 12/14/22 16:21:41.437 -Dec 14 16:21:41.963: INFO: Successfully updated pod "adopt-release-h4dgz" -STEP: Checking that the Job readopts the Pod 12/14/22 16:21:41.964 -Dec 14 16:21:41.964: INFO: Waiting up to 15m0s for pod "adopt-release-h4dgz" in namespace "job-9907" to be "adopted" -Dec 14 16:21:41.974: INFO: Pod "adopt-release-h4dgz": Phase="Running", Reason="", readiness=true. Elapsed: 10.385709ms -Dec 14 16:21:43.982: INFO: Pod "adopt-release-h4dgz": Phase="Running", Reason="", readiness=true. Elapsed: 2.017865552s -Dec 14 16:21:43.982: INFO: Pod "adopt-release-h4dgz" satisfied condition "adopted" -STEP: Removing the labels from the Job's Pod 12/14/22 16:21:43.982 -Dec 14 16:21:44.514: INFO: Successfully updated pod "adopt-release-h4dgz" -STEP: Checking that the Job releases the Pod 12/14/22 16:21:44.514 -Dec 14 16:21:44.515: INFO: Waiting up to 15m0s for pod "adopt-release-h4dgz" in namespace "job-9907" to be "released" -Dec 14 16:21:44.521: INFO: Pod "adopt-release-h4dgz": Phase="Running", Reason="", readiness=true. Elapsed: 6.27645ms -Dec 14 16:21:46.534: INFO: Pod "adopt-release-h4dgz": Phase="Running", Reason="", readiness=true. Elapsed: 2.018734426s -Dec 14 16:21:46.534: INFO: Pod "adopt-release-h4dgz" satisfied condition "released" -[AfterEach] [sig-apps] Job +STEP: Creating a kubernetes client 02/25/23 10:17:27.317 +Feb 25 10:17:27.318: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename gc 02/25/23 10:17:27.32 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:17:27.368 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:17:27.374 +[It] should not be blocked by dependency circle [Conformance] + test/e2e/apimachinery/garbage_collector.go:849 +Feb 25 10:17:27.467: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"71e9ff5b-73bf-482b-b0b0-571b4f0ffc6a", Controller:(*bool)(0xc00460f6be), BlockOwnerDeletion:(*bool)(0xc00460f6bf)}} +Feb 25 10:17:27.497: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"ec05f377-fbf3-44cd-a3dc-e7c77f89795a", Controller:(*bool)(0xc00460f8f6), BlockOwnerDeletion:(*bool)(0xc00460f8f7)}} +Feb 25 10:17:27.512: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"511bc339-57f5-4fe9-aa64-612459d3c346", Controller:(*bool)(0xc00460fb0a), BlockOwnerDeletion:(*bool)(0xc00460fb0b)}} +[AfterEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:187 -Dec 14 16:21:46.534: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "job-9907" for this suite. 12/14/22 16:21:46.54 -{"msg":"PASSED [sig-apps] Job should adopt matching orphans and release non-matching pods [Conformance]","completed":121,"skipped":2209,"failed":0} +Feb 25 10:17:32.540: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "gc-3781" for this suite. 02/25/23 10:17:32.549 +{"msg":"PASSED [sig-api-machinery] Garbage collector should not be blocked by dependency circle [Conformance]","completed":114,"skipped":2114,"failed":0} ------------------------------ -• [SLOW TEST] [7.166 seconds] -[sig-apps] Job -test/e2e/apps/framework.go:23 - should adopt matching orphans and release non-matching pods [Conformance] - test/e2e/apps/job.go:335 +• [SLOW TEST] [5.249 seconds] +[sig-api-machinery] Garbage collector +test/e2e/apimachinery/framework.go:23 + should not be blocked by dependency circle [Conformance] + test/e2e/apimachinery/garbage_collector.go:849 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Job + [BeforeEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:21:39.387 - Dec 14 16:21:39.388: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename job 12/14/22 16:21:39.392 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:39.412 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:39.417 - [It] should adopt matching orphans and release non-matching pods [Conformance] - test/e2e/apps/job.go:335 - STEP: Creating a job 12/14/22 16:21:39.42 - STEP: Ensuring active pods == parallelism 12/14/22 16:21:39.429 - STEP: Orphaning one of the Job's Pods 12/14/22 16:21:41.437 - Dec 14 16:21:41.963: INFO: Successfully updated pod "adopt-release-h4dgz" - STEP: Checking that the Job readopts the Pod 12/14/22 16:21:41.964 - Dec 14 16:21:41.964: INFO: Waiting up to 15m0s for pod "adopt-release-h4dgz" in namespace "job-9907" to be "adopted" - Dec 14 16:21:41.974: INFO: Pod "adopt-release-h4dgz": Phase="Running", Reason="", readiness=true. Elapsed: 10.385709ms - Dec 14 16:21:43.982: INFO: Pod "adopt-release-h4dgz": Phase="Running", Reason="", readiness=true. Elapsed: 2.017865552s - Dec 14 16:21:43.982: INFO: Pod "adopt-release-h4dgz" satisfied condition "adopted" - STEP: Removing the labels from the Job's Pod 12/14/22 16:21:43.982 - Dec 14 16:21:44.514: INFO: Successfully updated pod "adopt-release-h4dgz" - STEP: Checking that the Job releases the Pod 12/14/22 16:21:44.514 - Dec 14 16:21:44.515: INFO: Waiting up to 15m0s for pod "adopt-release-h4dgz" in namespace "job-9907" to be "released" - Dec 14 16:21:44.521: INFO: Pod "adopt-release-h4dgz": Phase="Running", Reason="", readiness=true. Elapsed: 6.27645ms - Dec 14 16:21:46.534: INFO: Pod "adopt-release-h4dgz": Phase="Running", Reason="", readiness=true. Elapsed: 2.018734426s - Dec 14 16:21:46.534: INFO: Pod "adopt-release-h4dgz" satisfied condition "released" - [AfterEach] [sig-apps] Job + STEP: Creating a kubernetes client 02/25/23 10:17:27.317 + Feb 25 10:17:27.318: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename gc 02/25/23 10:17:27.32 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:17:27.368 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:17:27.374 + [It] should not be blocked by dependency circle [Conformance] + test/e2e/apimachinery/garbage_collector.go:849 + Feb 25 10:17:27.467: INFO: pod1.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod3", UID:"71e9ff5b-73bf-482b-b0b0-571b4f0ffc6a", Controller:(*bool)(0xc00460f6be), BlockOwnerDeletion:(*bool)(0xc00460f6bf)}} + Feb 25 10:17:27.497: INFO: pod2.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod1", UID:"ec05f377-fbf3-44cd-a3dc-e7c77f89795a", Controller:(*bool)(0xc00460f8f6), BlockOwnerDeletion:(*bool)(0xc00460f8f7)}} + Feb 25 10:17:27.512: INFO: pod3.ObjectMeta.OwnerReferences=[]v1.OwnerReference{v1.OwnerReference{APIVersion:"v1", Kind:"Pod", Name:"pod2", UID:"511bc339-57f5-4fe9-aa64-612459d3c346", Controller:(*bool)(0xc00460fb0a), BlockOwnerDeletion:(*bool)(0xc00460fb0b)}} + [AfterEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:187 - Dec 14 16:21:46.534: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "job-9907" for this suite. 12/14/22 16:21:46.54 + Feb 25 10:17:32.540: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "gc-3781" for this suite. 02/25/23 10:17:32.549 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Secrets - should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:56 -[BeforeEach] [sig-storage] Secrets +[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] + should perform rolling updates and roll backs of template modifications [Conformance] + test/e2e/apps/statefulset.go:304 +[BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:21:46.559 -Dec 14 16:21:46.559: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename secrets 12/14/22 16:21:46.561 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:46.581 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:46.584 -[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:56 -STEP: Creating secret with name secret-test-6c656dbd-0663-400f-a981-5a2af5d082e1 12/14/22 16:21:46.587 -STEP: Creating a pod to test consume secrets 12/14/22 16:21:46.594 -Dec 14 16:21:46.605: INFO: Waiting up to 5m0s for pod "pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517" in namespace "secrets-5625" to be "Succeeded or Failed" -Dec 14 16:21:46.615: INFO: Pod "pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517": Phase="Pending", Reason="", readiness=false. Elapsed: 9.946303ms -Dec 14 16:21:48.625: INFO: Pod "pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517": Phase="Running", Reason="", readiness=false. Elapsed: 2.01960604s -Dec 14 16:21:50.624: INFO: Pod "pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018966647s -STEP: Saw pod success 12/14/22 16:21:50.624 -Dec 14 16:21:50.624: INFO: Pod "pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517" satisfied condition "Succeeded or Failed" -Dec 14 16:21:50.630: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517 container secret-volume-test: -STEP: delete the pod 12/14/22 16:21:50.643 -Dec 14 16:21:50.659: INFO: Waiting for pod pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517 to disappear -Dec 14 16:21:50.663: INFO: Pod pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517 no longer exists -[AfterEach] [sig-storage] Secrets +STEP: Creating a kubernetes client 02/25/23 10:17:32.577 +Feb 25 10:17:32.577: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename statefulset 02/25/23 10:17:32.58 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:17:32.617 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:17:32.625 +[BeforeEach] [sig-apps] StatefulSet + test/e2e/apps/statefulset.go:96 +[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:111 +STEP: Creating service test in namespace statefulset-4315 02/25/23 10:17:32.639 +[It] should perform rolling updates and roll backs of template modifications [Conformance] + test/e2e/apps/statefulset.go:304 +STEP: Creating a new StatefulSet 02/25/23 10:17:32.652 +Feb 25 10:17:32.685: INFO: Found 0 stateful pods, waiting for 3 +Feb 25 10:17:42.695: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true +Feb 25 10:17:42.695: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true +Feb 25 10:17:42.695: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true +Feb 25 10:17:42.719: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-4315 exec ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Feb 25 10:17:43.136: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Feb 25 10:17:43.136: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Feb 25 10:17:43.136: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +STEP: Updating StatefulSet template: update image from registry.k8s.io/e2e-test-images/httpd:2.4.38-2 to registry.k8s.io/e2e-test-images/httpd:2.4.39-2 02/25/23 10:17:53.171 +Feb 25 10:17:53.207: INFO: Updating stateful set ss2 +STEP: Creating a new revision 02/25/23 10:17:53.207 +STEP: Updating Pods in reverse ordinal order 02/25/23 10:18:03.242 +Feb 25 10:18:03.251: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-4315 exec ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' +Feb 25 10:18:03.598: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" +Feb 25 10:18:03.598: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" +Feb 25 10:18:03.598: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + +Feb 25 10:18:13.661: INFO: Waiting for StatefulSet statefulset-4315/ss2 to complete update +Feb 25 10:18:13.661: INFO: Waiting for Pod statefulset-4315/ss2-0 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 +Feb 25 10:18:13.662: INFO: Waiting for Pod statefulset-4315/ss2-1 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 +Feb 25 10:18:23.682: INFO: Waiting for StatefulSet statefulset-4315/ss2 to complete update +Feb 25 10:18:23.682: INFO: Waiting for Pod statefulset-4315/ss2-0 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 +Feb 25 10:18:33.683: INFO: Waiting for StatefulSet statefulset-4315/ss2 to complete update +Feb 25 10:18:43.686: INFO: Waiting for StatefulSet statefulset-4315/ss2 to complete update +STEP: Rolling back to a previous revision 02/25/23 10:18:53.681 +Feb 25 10:18:53.682: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-4315 exec ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Feb 25 10:18:54.105: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Feb 25 10:18:54.105: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Feb 25 10:18:54.105: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +Feb 25 10:19:04.178: INFO: Updating stateful set ss2 +STEP: Rolling back update in reverse ordinal order 02/25/23 10:19:14.217 +Feb 25 10:19:14.226: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-4315 exec ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' +Feb 25 10:19:14.539: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" +Feb 25 10:19:14.540: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" +Feb 25 10:19:14.540: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + +[AfterEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:122 +Feb 25 10:19:24.595: INFO: Deleting all statefulset in ns statefulset-4315 +Feb 25 10:19:24.603: INFO: Scaling statefulset ss2 to 0 +Feb 25 10:19:34.650: INFO: Waiting for statefulset status.replicas updated to 0 +Feb 25 10:19:34.658: INFO: Deleting statefulset ss2 +[AfterEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:187 -Dec 14 16:21:50.664: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-5625" for this suite. 12/14/22 16:21:50.671 -{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","completed":122,"skipped":2233,"failed":0} +Feb 25 10:19:34.706: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "statefulset-4315" for this suite. 02/25/23 10:19:34.719 +{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform rolling updates and roll backs of template modifications [Conformance]","completed":115,"skipped":2145,"failed":0} ------------------------------ -• [4.125 seconds] -[sig-storage] Secrets -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:56 +• [SLOW TEST] [122.161 seconds] +[sig-apps] StatefulSet +test/e2e/apps/framework.go:23 + Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:101 + should perform rolling updates and roll backs of template modifications [Conformance] + test/e2e/apps/statefulset.go:304 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Secrets + [BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:21:46.559 - Dec 14 16:21:46.559: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename secrets 12/14/22 16:21:46.561 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:46.581 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:46.584 - [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:56 - STEP: Creating secret with name secret-test-6c656dbd-0663-400f-a981-5a2af5d082e1 12/14/22 16:21:46.587 - STEP: Creating a pod to test consume secrets 12/14/22 16:21:46.594 - Dec 14 16:21:46.605: INFO: Waiting up to 5m0s for pod "pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517" in namespace "secrets-5625" to be "Succeeded or Failed" - Dec 14 16:21:46.615: INFO: Pod "pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517": Phase="Pending", Reason="", readiness=false. Elapsed: 9.946303ms - Dec 14 16:21:48.625: INFO: Pod "pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517": Phase="Running", Reason="", readiness=false. Elapsed: 2.01960604s - Dec 14 16:21:50.624: INFO: Pod "pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018966647s - STEP: Saw pod success 12/14/22 16:21:50.624 - Dec 14 16:21:50.624: INFO: Pod "pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517" satisfied condition "Succeeded or Failed" - Dec 14 16:21:50.630: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517 container secret-volume-test: - STEP: delete the pod 12/14/22 16:21:50.643 - Dec 14 16:21:50.659: INFO: Waiting for pod pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517 to disappear - Dec 14 16:21:50.663: INFO: Pod pod-secrets-b10e8267-91c2-48da-8f58-9002b38aa517 no longer exists - [AfterEach] [sig-storage] Secrets + STEP: Creating a kubernetes client 02/25/23 10:17:32.577 + Feb 25 10:17:32.577: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename statefulset 02/25/23 10:17:32.58 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:17:32.617 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:17:32.625 + [BeforeEach] [sig-apps] StatefulSet + test/e2e/apps/statefulset.go:96 + [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:111 + STEP: Creating service test in namespace statefulset-4315 02/25/23 10:17:32.639 + [It] should perform rolling updates and roll backs of template modifications [Conformance] + test/e2e/apps/statefulset.go:304 + STEP: Creating a new StatefulSet 02/25/23 10:17:32.652 + Feb 25 10:17:32.685: INFO: Found 0 stateful pods, waiting for 3 + Feb 25 10:17:42.695: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true + Feb 25 10:17:42.695: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true + Feb 25 10:17:42.695: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true + Feb 25 10:17:42.719: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-4315 exec ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' + Feb 25 10:17:43.136: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" + Feb 25 10:17:43.136: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" + Feb 25 10:17:43.136: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + + STEP: Updating StatefulSet template: update image from registry.k8s.io/e2e-test-images/httpd:2.4.38-2 to registry.k8s.io/e2e-test-images/httpd:2.4.39-2 02/25/23 10:17:53.171 + Feb 25 10:17:53.207: INFO: Updating stateful set ss2 + STEP: Creating a new revision 02/25/23 10:17:53.207 + STEP: Updating Pods in reverse ordinal order 02/25/23 10:18:03.242 + Feb 25 10:18:03.251: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-4315 exec ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' + Feb 25 10:18:03.598: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" + Feb 25 10:18:03.598: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" + Feb 25 10:18:03.598: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + + Feb 25 10:18:13.661: INFO: Waiting for StatefulSet statefulset-4315/ss2 to complete update + Feb 25 10:18:13.661: INFO: Waiting for Pod statefulset-4315/ss2-0 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 + Feb 25 10:18:13.662: INFO: Waiting for Pod statefulset-4315/ss2-1 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 + Feb 25 10:18:23.682: INFO: Waiting for StatefulSet statefulset-4315/ss2 to complete update + Feb 25 10:18:23.682: INFO: Waiting for Pod statefulset-4315/ss2-0 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 + Feb 25 10:18:33.683: INFO: Waiting for StatefulSet statefulset-4315/ss2 to complete update + Feb 25 10:18:43.686: INFO: Waiting for StatefulSet statefulset-4315/ss2 to complete update + STEP: Rolling back to a previous revision 02/25/23 10:18:53.681 + Feb 25 10:18:53.682: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-4315 exec ss2-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' + Feb 25 10:18:54.105: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" + Feb 25 10:18:54.105: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" + Feb 25 10:18:54.105: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss2-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + + Feb 25 10:19:04.178: INFO: Updating stateful set ss2 + STEP: Rolling back update in reverse ordinal order 02/25/23 10:19:14.217 + Feb 25 10:19:14.226: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-4315 exec ss2-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' + Feb 25 10:19:14.539: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" + Feb 25 10:19:14.540: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" + Feb 25 10:19:14.540: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss2-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + + [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:122 + Feb 25 10:19:24.595: INFO: Deleting all statefulset in ns statefulset-4315 + Feb 25 10:19:24.603: INFO: Scaling statefulset ss2 to 0 + Feb 25 10:19:34.650: INFO: Waiting for statefulset status.replicas updated to 0 + Feb 25 10:19:34.658: INFO: Deleting statefulset ss2 + [AfterEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:187 - Dec 14 16:21:50.664: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "secrets-5625" for this suite. 12/14/22 16:21:50.671 + Feb 25 10:19:34.706: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "statefulset-4315" for this suite. 02/25/23 10:19:34.719 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- [sig-storage] Secrets - should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:98 + should be immutable if `immutable` field is set [Conformance] + test/e2e/common/storage/secrets_volume.go:385 [BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:21:50.69 -Dec 14 16:21:50.690: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename secrets 12/14/22 16:21:50.693 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:50.716 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:50.722 -[It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:98 -STEP: Creating secret with name secret-test-da9cf16d-0bc1-4648-bfa5-8e9ec42e6610 12/14/22 16:21:50.765 -STEP: Creating a pod to test consume secrets 12/14/22 16:21:50.774 -Dec 14 16:21:50.786: INFO: Waiting up to 5m0s for pod "pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401" in namespace "secrets-5470" to be "Succeeded or Failed" -Dec 14 16:21:50.792: INFO: Pod "pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401": Phase="Pending", Reason="", readiness=false. Elapsed: 6.735757ms -Dec 14 16:21:52.802: INFO: Pod "pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016107739s -Dec 14 16:21:54.803: INFO: Pod "pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.017041527s -STEP: Saw pod success 12/14/22 16:21:54.803 -Dec 14 16:21:54.803: INFO: Pod "pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401" satisfied condition "Succeeded or Failed" -Dec 14 16:21:54.808: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401 container secret-volume-test: -STEP: delete the pod 12/14/22 16:21:54.816 -Dec 14 16:21:54.831: INFO: Waiting for pod pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401 to disappear -Dec 14 16:21:54.834: INFO: Pod pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401 no longer exists +STEP: Creating a kubernetes client 02/25/23 10:19:34.739 +Feb 25 10:19:34.739: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename secrets 02/25/23 10:19:34.745 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:19:34.806 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:19:34.813 +[It] should be immutable if `immutable` field is set [Conformance] + test/e2e/common/storage/secrets_volume.go:385 [AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 -Dec 14 16:21:54.835: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-5470" for this suite. 12/14/22 16:21:54.841 -STEP: Destroying namespace "secret-namespace-3406" for this suite. 12/14/22 16:21:54.847 -{"msg":"PASSED [sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]","completed":123,"skipped":2261,"failed":0} +Feb 25 10:19:34.945: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-3182" for this suite. 02/25/23 10:19:34.952 +{"msg":"PASSED [sig-storage] Secrets should be immutable if `immutable` field is set [Conformance]","completed":116,"skipped":2145,"failed":0} ------------------------------ -• [4.169 seconds] +• [0.230 seconds] [sig-storage] Secrets test/e2e/common/storage/framework.go:23 - should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:98 + should be immutable if `immutable` field is set [Conformance] + test/e2e/common/storage/secrets_volume.go:385 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:21:50.69 - Dec 14 16:21:50.690: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename secrets 12/14/22 16:21:50.693 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:50.716 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:50.722 - [It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:98 - STEP: Creating secret with name secret-test-da9cf16d-0bc1-4648-bfa5-8e9ec42e6610 12/14/22 16:21:50.765 - STEP: Creating a pod to test consume secrets 12/14/22 16:21:50.774 - Dec 14 16:21:50.786: INFO: Waiting up to 5m0s for pod "pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401" in namespace "secrets-5470" to be "Succeeded or Failed" - Dec 14 16:21:50.792: INFO: Pod "pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401": Phase="Pending", Reason="", readiness=false. Elapsed: 6.735757ms - Dec 14 16:21:52.802: INFO: Pod "pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016107739s - Dec 14 16:21:54.803: INFO: Pod "pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.017041527s - STEP: Saw pod success 12/14/22 16:21:54.803 - Dec 14 16:21:54.803: INFO: Pod "pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401" satisfied condition "Succeeded or Failed" - Dec 14 16:21:54.808: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401 container secret-volume-test: - STEP: delete the pod 12/14/22 16:21:54.816 - Dec 14 16:21:54.831: INFO: Waiting for pod pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401 to disappear - Dec 14 16:21:54.834: INFO: Pod pod-secrets-0459cfe6-cad0-411a-87ed-b5f0058c6401 no longer exists + STEP: Creating a kubernetes client 02/25/23 10:19:34.739 + Feb 25 10:19:34.739: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename secrets 02/25/23 10:19:34.745 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:19:34.806 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:19:34.813 + [It] should be immutable if `immutable` field is set [Conformance] + test/e2e/common/storage/secrets_volume.go:385 [AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 - Dec 14 16:21:54.835: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "secrets-5470" for this suite. 12/14/22 16:21:54.841 - STEP: Destroying namespace "secret-namespace-3406" for this suite. 12/14/22 16:21:54.847 + Feb 25 10:19:34.945: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "secrets-3182" for this suite. 02/25/23 10:19:34.952 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSS +SSSSSSSS ------------------------------ -[sig-apps] Daemon set [Serial] - should list and delete a collection of DaemonSets [Conformance] - test/e2e/apps/daemon_set.go:822 -[BeforeEach] [sig-apps] Daemon set [Serial] +[sig-network] Services + should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance] + test/e2e/network/service.go:2157 +[BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:21:54.86 -Dec 14 16:21:54.861: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename daemonsets 12/14/22 16:21:54.863 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:54.882 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:54.886 -[BeforeEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:145 -[It] should list and delete a collection of DaemonSets [Conformance] - test/e2e/apps/daemon_set.go:822 -STEP: Creating simple DaemonSet "daemon-set" 12/14/22 16:21:54.925 -STEP: Check that daemon pods launch on every node of the cluster. 12/14/22 16:21:54.935 -Dec 14 16:21:54.948: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:21:54.948: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 16:21:55.960: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:21:55.960: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 16:21:56.970: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 -Dec 14 16:21:56.970: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set -STEP: listing all DeamonSets 12/14/22 16:21:56.977 -STEP: DeleteCollection of the DaemonSets 12/14/22 16:21:56.983 -STEP: Verify that ReplicaSets have been deleted 12/14/22 16:21:56.999 -[AfterEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:110 -Dec 14 16:21:57.027: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"15997"},"items":null} - -Dec 14 16:21:57.042: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"15999"},"items":[{"metadata":{"name":"daemon-set-cmkrw","generateName":"daemon-set-","namespace":"daemonsets-1968","uid":"91a7b936-7392-4e66-bd28-f4461c5f25ef","resourceVersion":"15999","creationTimestamp":"2022-12-14T16:21:54Z","deletionTimestamp":"2022-12-14T16:22:27Z","deletionGracePeriodSeconds":30,"labels":{"controller-revision-hash":"7f7ffb4fcc","daemonset-name":"daemon-set","pod-template-generation":"1"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"DaemonSet","name":"daemon-set","uid":"4be153d1-be77-4031-b20e-73937d0ff050","controller":true,"blockOwnerDeletion":true}],"managedFields":[{"manager":"kube-controller-manager","operation":"Update","apiVersion":"v1","time":"2022-12-14T16:21:54Z","fieldsType":"FieldsV1","fieldsV1":{"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:controller-revision-hash":{},"f:daemonset-name":{},"f:pod-template-generation":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4be153d1-be77-4031-b20e-73937d0ff050\"}":{}}},"f:spec":{"f:affinity":{".":{},"f:nodeAffinity":{".":{},"f:requiredDuringSchedulingIgnoredDuringExecution":{}}},"f:containers":{"k:{\"name\":\"app\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:ports":{".":{},"k:{\"containerPort\":9376,\"protocol\":\"TCP\"}":{".":{},"f:containerPort":{},"f:protocol":{}}},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{},"f:tolerations":{}}}},{"manager":"kubelet","operation":"Update","apiVersion":"v1","time":"2022-12-14T16:21:56Z","fieldsType":"FieldsV1","fieldsV1":{"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.205\"}":{".":{},"f:ip":{}}},"f:startTime":{}}},"subresource":"status"}]},"spec":{"volumes":[{"name":"kube-api-access-mvrtz","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"app","image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","ports":[{"containerPort":9376,"protocol":"TCP"}],"resources":{},"volumeMounts":[{"name":"kube-api-access-mvrtz","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"File","imagePullPolicy":"IfNotPresent","securityContext":{}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","serviceAccountName":"default","serviceAccount":"default","nodeName":"kiedai4iakae-3","securityContext":{},"affinity":{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchFields":[{"key":"metadata.name","operator":"In","values":["kiedai4iakae-3"]}]}]}}},"schedulerName":"default-scheduler","tolerations":[{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/disk-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/pid-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unschedulable","operator":"Exists","effect":"NoSchedule"}],"priority":0,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:54Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:56Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:56Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:54Z"}],"hostIP":"192.168.121.244","podIP":"10.233.66.205","podIPs":[{"ip":"10.233.66.205"}],"startTime":"2022-12-14T16:21:54Z","containerStatuses":[{"name":"app","state":{"running":{"startedAt":"2022-12-14T16:21:55Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","imageID":"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3","containerID":"cri-o://dbedb98a5fc7c0a783b7af9397523466a97decf06f8f180f360b9f2883dfbc1c","started":true}],"qosClass":"BestEffort"}},{"metadata":{"name":"daemon-set-ndp8d","generateName":"daemon-set-","namespace":"daemonsets-1968","uid":"368dbf45-3464-4d42-acc2-a97c02adcb24","resourceVersion":"15997","creationTimestamp":"2022-12-14T16:21:54Z","deletionTimestamp":"2022-12-14T16:22:26Z","deletionGracePeriodSeconds":30,"labels":{"controller-revision-hash":"7f7ffb4fcc","daemonset-name":"daemon-set","pod-template-generation":"1"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"DaemonSet","name":"daemon-set","uid":"4be153d1-be77-4031-b20e-73937d0ff050","controller":true,"blockOwnerDeletion":true}],"managedFields":[{"manager":"kube-controller-manager","operation":"Update","apiVersion":"v1","time":"2022-12-14T16:21:54Z","fieldsType":"FieldsV1","fieldsV1":{"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:controller-revision-hash":{},"f:daemonset-name":{},"f:pod-template-generation":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4be153d1-be77-4031-b20e-73937d0ff050\"}":{}}},"f:spec":{"f:affinity":{".":{},"f:nodeAffinity":{".":{},"f:requiredDuringSchedulingIgnoredDuringExecution":{}}},"f:containers":{"k:{\"name\":\"app\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:ports":{".":{},"k:{\"containerPort\":9376,\"protocol\":\"TCP\"}":{".":{},"f:containerPort":{},"f:protocol":{}}},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{},"f:tolerations":{}}}},{"manager":"kubelet","operation":"Update","apiVersion":"v1","time":"2022-12-14T16:21:56Z","fieldsType":"FieldsV1","fieldsV1":{"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.96\"}":{".":{},"f:ip":{}}},"f:startTime":{}}},"subresource":"status"}]},"spec":{"volumes":[{"name":"kube-api-access-9lxft","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"app","image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","ports":[{"containerPort":9376,"protocol":"TCP"}],"resources":{},"volumeMounts":[{"name":"kube-api-access-9lxft","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"File","imagePullPolicy":"IfNotPresent","securityContext":{}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","serviceAccountName":"default","serviceAccount":"default","nodeName":"kiedai4iakae-1","securityContext":{},"affinity":{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchFields":[{"key":"metadata.name","operator":"In","values":["kiedai4iakae-1"]}]}]}}},"schedulerName":"default-scheduler","tolerations":[{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/disk-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/pid-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unschedulable","operator":"Exists","effect":"NoSchedule"}],"priority":0,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:54Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:56Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:56Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:54Z"}],"hostIP":"192.168.121.249","podIP":"10.233.64.96","podIPs":[{"ip":"10.233.64.96"}],"startTime":"2022-12-14T16:21:54Z","containerStatuses":[{"name":"app","state":{"running":{"startedAt":"2022-12-14T16:21:56Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","imageID":"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3","containerID":"cri-o://7ed1cbe4c335b4c061d02e7f3085d63253e5126e7fc591d501ac9895e5f0fead","started":true}],"qosClass":"BestEffort"}},{"metadata":{"name":"daemon-set-rlpvz","generateName":"daemon-set-","namespace":"daemonsets-1968","uid":"eb208fdd-0302-4b90-a61a-985946288a62","resourceVersion":"15996","creationTimestamp":"2022-12-14T16:21:54Z","deletionTimestamp":"2022-12-14T16:22:26Z","deletionGracePeriodSeconds":30,"labels":{"controller-revision-hash":"7f7ffb4fcc","daemonset-name":"daemon-set","pod-template-generation":"1"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"DaemonSet","name":"daemon-set","uid":"4be153d1-be77-4031-b20e-73937d0ff050","controller":true,"blockOwnerDeletion":true}],"managedFields":[{"manager":"kube-controller-manager","operation":"Update","apiVersion":"v1","time":"2022-12-14T16:21:54Z","fieldsType":"FieldsV1","fieldsV1":{"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:controller-revision-hash":{},"f:daemonset-name":{},"f:pod-template-generation":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4be153d1-be77-4031-b20e-73937d0ff050\"}":{}}},"f:spec":{"f:affinity":{".":{},"f:nodeAffinity":{".":{},"f:requiredDuringSchedulingIgnoredDuringExecution":{}}},"f:containers":{"k:{\"name\":\"app\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:ports":{".":{},"k:{\"containerPort\":9376,\"protocol\":\"TCP\"}":{".":{},"f:containerPort":{},"f:protocol":{}}},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{},"f:tolerations":{}}}},{"manager":"kubelet","operation":"Update","apiVersion":"v1","time":"2022-12-14T16:21:56Z","fieldsType":"FieldsV1","fieldsV1":{"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.93\"}":{".":{},"f:ip":{}}},"f:startTime":{}}},"subresource":"status"}]},"spec":{"volumes":[{"name":"kube-api-access-95z2g","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"app","image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","ports":[{"containerPort":9376,"protocol":"TCP"}],"resources":{},"volumeMounts":[{"name":"kube-api-access-95z2g","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"File","imagePullPolicy":"IfNotPresent","securityContext":{}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","serviceAccountName":"default","serviceAccount":"default","nodeName":"kiedai4iakae-2","securityContext":{},"affinity":{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchFields":[{"key":"metadata.name","operator":"In","values":["kiedai4iakae-2"]}]}]}}},"schedulerName":"default-scheduler","tolerations":[{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/disk-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/pid-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unschedulable","operator":"Exists","effect":"NoSchedule"}],"priority":0,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:54Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:56Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:56Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:54Z"}],"hostIP":"192.168.121.197","podIP":"10.233.65.93","podIPs":[{"ip":"10.233.65.93"}],"startTime":"2022-12-14T16:21:54Z","containerStatuses":[{"name":"app","state":{"running":{"startedAt":"2022-12-14T16:21:55Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","imageID":"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3","containerID":"cri-o://94d323b9905a77ef322ce7e37b1c8f77e2d36bb68acca8f12b214b91ed17b4f3","started":true}],"qosClass":"BestEffort"}}]} - -[AfterEach] [sig-apps] Daemon set [Serial] +STEP: Creating a kubernetes client 02/25/23 10:19:34.97 +Feb 25 10:19:34.971: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 10:19:34.972 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:19:35.008 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:19:35.015 +[BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 +[It] should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance] + test/e2e/network/service.go:2157 +STEP: creating service in namespace services-542 02/25/23 10:19:35.022 +STEP: creating service affinity-clusterip in namespace services-542 02/25/23 10:19:35.022 +STEP: creating replication controller affinity-clusterip in namespace services-542 02/25/23 10:19:35.052 +I0225 10:19:35.069143 13 runners.go:193] Created replication controller with name: affinity-clusterip, namespace: services-542, replica count: 3 +I0225 10:19:38.121936 13 runners.go:193] affinity-clusterip Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Feb 25 10:19:38.140: INFO: Creating new exec pod +Feb 25 10:19:38.160: INFO: Waiting up to 5m0s for pod "execpod-affinityxwzhq" in namespace "services-542" to be "running" +Feb 25 10:19:38.170: INFO: Pod "execpod-affinityxwzhq": Phase="Pending", Reason="", readiness=false. Elapsed: 9.091165ms +Feb 25 10:19:40.182: INFO: Pod "execpod-affinityxwzhq": Phase="Running", Reason="", readiness=true. Elapsed: 2.021654156s +Feb 25 10:19:40.182: INFO: Pod "execpod-affinityxwzhq" satisfied condition "running" +Feb 25 10:19:41.184: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-542 exec execpod-affinityxwzhq -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-clusterip 80' +Feb 25 10:19:41.589: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-clusterip 80\nConnection to affinity-clusterip 80 port [tcp/http] succeeded!\n" +Feb 25 10:19:41.589: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 10:19:41.589: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-542 exec execpod-affinityxwzhq -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.32.118 80' +Feb 25 10:19:42.073: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.32.118 80\nConnection to 10.233.32.118 80 port [tcp/http] succeeded!\n" +Feb 25 10:19:42.073: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 10:19:42.074: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-542 exec execpod-affinityxwzhq -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.32.118:80/ ; done' +Feb 25 10:19:42.808: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n" +Feb 25 10:19:42.808: INFO: stdout: "\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9" +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 +Feb 25 10:19:42.808: INFO: Cleaning up the exec pod +STEP: deleting ReplicationController affinity-clusterip in namespace services-542, will wait for the garbage collector to delete the pods 02/25/23 10:19:42.847 +Feb 25 10:19:42.927: INFO: Deleting ReplicationController affinity-clusterip took: 16.734723ms +Feb 25 10:19:43.029: INFO: Terminating ReplicationController affinity-clusterip pods took: 102.194232ms +[AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 16:21:57.063: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "daemonsets-1968" for this suite. 12/14/22 16:21:57.08 -{"msg":"PASSED [sig-apps] Daemon set [Serial] should list and delete a collection of DaemonSets [Conformance]","completed":124,"skipped":2275,"failed":0} +Feb 25 10:19:45.772: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-542" for this suite. 02/25/23 10:19:45.788 +[AfterEach] [sig-network] Services + test/e2e/network/service.go:762 +{"msg":"PASSED [sig-network] Services should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance]","completed":117,"skipped":2153,"failed":0} ------------------------------ -• [2.236 seconds] -[sig-apps] Daemon set [Serial] -test/e2e/apps/framework.go:23 - should list and delete a collection of DaemonSets [Conformance] - test/e2e/apps/daemon_set.go:822 +• [SLOW TEST] [10.834 seconds] +[sig-network] Services +test/e2e/network/common/framework.go:23 + should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance] + test/e2e/network/service.go:2157 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Daemon set [Serial] + [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:21:54.86 - Dec 14 16:21:54.861: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename daemonsets 12/14/22 16:21:54.863 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:54.882 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:54.886 - [BeforeEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:145 - [It] should list and delete a collection of DaemonSets [Conformance] - test/e2e/apps/daemon_set.go:822 - STEP: Creating simple DaemonSet "daemon-set" 12/14/22 16:21:54.925 - STEP: Check that daemon pods launch on every node of the cluster. 12/14/22 16:21:54.935 - Dec 14 16:21:54.948: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:21:54.948: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 16:21:55.960: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:21:55.960: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 16:21:56.970: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 - Dec 14 16:21:56.970: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set - STEP: listing all DeamonSets 12/14/22 16:21:56.977 - STEP: DeleteCollection of the DaemonSets 12/14/22 16:21:56.983 - STEP: Verify that ReplicaSets have been deleted 12/14/22 16:21:56.999 - [AfterEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:110 - Dec 14 16:21:57.027: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"15997"},"items":null} - - Dec 14 16:21:57.042: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"15999"},"items":[{"metadata":{"name":"daemon-set-cmkrw","generateName":"daemon-set-","namespace":"daemonsets-1968","uid":"91a7b936-7392-4e66-bd28-f4461c5f25ef","resourceVersion":"15999","creationTimestamp":"2022-12-14T16:21:54Z","deletionTimestamp":"2022-12-14T16:22:27Z","deletionGracePeriodSeconds":30,"labels":{"controller-revision-hash":"7f7ffb4fcc","daemonset-name":"daemon-set","pod-template-generation":"1"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"DaemonSet","name":"daemon-set","uid":"4be153d1-be77-4031-b20e-73937d0ff050","controller":true,"blockOwnerDeletion":true}],"managedFields":[{"manager":"kube-controller-manager","operation":"Update","apiVersion":"v1","time":"2022-12-14T16:21:54Z","fieldsType":"FieldsV1","fieldsV1":{"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:controller-revision-hash":{},"f:daemonset-name":{},"f:pod-template-generation":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4be153d1-be77-4031-b20e-73937d0ff050\"}":{}}},"f:spec":{"f:affinity":{".":{},"f:nodeAffinity":{".":{},"f:requiredDuringSchedulingIgnoredDuringExecution":{}}},"f:containers":{"k:{\"name\":\"app\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:ports":{".":{},"k:{\"containerPort\":9376,\"protocol\":\"TCP\"}":{".":{},"f:containerPort":{},"f:protocol":{}}},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{},"f:tolerations":{}}}},{"manager":"kubelet","operation":"Update","apiVersion":"v1","time":"2022-12-14T16:21:56Z","fieldsType":"FieldsV1","fieldsV1":{"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.205\"}":{".":{},"f:ip":{}}},"f:startTime":{}}},"subresource":"status"}]},"spec":{"volumes":[{"name":"kube-api-access-mvrtz","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"app","image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","ports":[{"containerPort":9376,"protocol":"TCP"}],"resources":{},"volumeMounts":[{"name":"kube-api-access-mvrtz","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"File","imagePullPolicy":"IfNotPresent","securityContext":{}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","serviceAccountName":"default","serviceAccount":"default","nodeName":"kiedai4iakae-3","securityContext":{},"affinity":{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchFields":[{"key":"metadata.name","operator":"In","values":["kiedai4iakae-3"]}]}]}}},"schedulerName":"default-scheduler","tolerations":[{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/disk-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/pid-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unschedulable","operator":"Exists","effect":"NoSchedule"}],"priority":0,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:54Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:56Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:56Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:54Z"}],"hostIP":"192.168.121.244","podIP":"10.233.66.205","podIPs":[{"ip":"10.233.66.205"}],"startTime":"2022-12-14T16:21:54Z","containerStatuses":[{"name":"app","state":{"running":{"startedAt":"2022-12-14T16:21:55Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","imageID":"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3","containerID":"cri-o://dbedb98a5fc7c0a783b7af9397523466a97decf06f8f180f360b9f2883dfbc1c","started":true}],"qosClass":"BestEffort"}},{"metadata":{"name":"daemon-set-ndp8d","generateName":"daemon-set-","namespace":"daemonsets-1968","uid":"368dbf45-3464-4d42-acc2-a97c02adcb24","resourceVersion":"15997","creationTimestamp":"2022-12-14T16:21:54Z","deletionTimestamp":"2022-12-14T16:22:26Z","deletionGracePeriodSeconds":30,"labels":{"controller-revision-hash":"7f7ffb4fcc","daemonset-name":"daemon-set","pod-template-generation":"1"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"DaemonSet","name":"daemon-set","uid":"4be153d1-be77-4031-b20e-73937d0ff050","controller":true,"blockOwnerDeletion":true}],"managedFields":[{"manager":"kube-controller-manager","operation":"Update","apiVersion":"v1","time":"2022-12-14T16:21:54Z","fieldsType":"FieldsV1","fieldsV1":{"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:controller-revision-hash":{},"f:daemonset-name":{},"f:pod-template-generation":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4be153d1-be77-4031-b20e-73937d0ff050\"}":{}}},"f:spec":{"f:affinity":{".":{},"f:nodeAffinity":{".":{},"f:requiredDuringSchedulingIgnoredDuringExecution":{}}},"f:containers":{"k:{\"name\":\"app\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:ports":{".":{},"k:{\"containerPort\":9376,\"protocol\":\"TCP\"}":{".":{},"f:containerPort":{},"f:protocol":{}}},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{},"f:tolerations":{}}}},{"manager":"kubelet","operation":"Update","apiVersion":"v1","time":"2022-12-14T16:21:56Z","fieldsType":"FieldsV1","fieldsV1":{"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.96\"}":{".":{},"f:ip":{}}},"f:startTime":{}}},"subresource":"status"}]},"spec":{"volumes":[{"name":"kube-api-access-9lxft","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"app","image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","ports":[{"containerPort":9376,"protocol":"TCP"}],"resources":{},"volumeMounts":[{"name":"kube-api-access-9lxft","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"File","imagePullPolicy":"IfNotPresent","securityContext":{}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","serviceAccountName":"default","serviceAccount":"default","nodeName":"kiedai4iakae-1","securityContext":{},"affinity":{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchFields":[{"key":"metadata.name","operator":"In","values":["kiedai4iakae-1"]}]}]}}},"schedulerName":"default-scheduler","tolerations":[{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/disk-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/pid-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unschedulable","operator":"Exists","effect":"NoSchedule"}],"priority":0,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:54Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:56Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:56Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:54Z"}],"hostIP":"192.168.121.249","podIP":"10.233.64.96","podIPs":[{"ip":"10.233.64.96"}],"startTime":"2022-12-14T16:21:54Z","containerStatuses":[{"name":"app","state":{"running":{"startedAt":"2022-12-14T16:21:56Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","imageID":"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3","containerID":"cri-o://7ed1cbe4c335b4c061d02e7f3085d63253e5126e7fc591d501ac9895e5f0fead","started":true}],"qosClass":"BestEffort"}},{"metadata":{"name":"daemon-set-rlpvz","generateName":"daemon-set-","namespace":"daemonsets-1968","uid":"eb208fdd-0302-4b90-a61a-985946288a62","resourceVersion":"15996","creationTimestamp":"2022-12-14T16:21:54Z","deletionTimestamp":"2022-12-14T16:22:26Z","deletionGracePeriodSeconds":30,"labels":{"controller-revision-hash":"7f7ffb4fcc","daemonset-name":"daemon-set","pod-template-generation":"1"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"DaemonSet","name":"daemon-set","uid":"4be153d1-be77-4031-b20e-73937d0ff050","controller":true,"blockOwnerDeletion":true}],"managedFields":[{"manager":"kube-controller-manager","operation":"Update","apiVersion":"v1","time":"2022-12-14T16:21:54Z","fieldsType":"FieldsV1","fieldsV1":{"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:controller-revision-hash":{},"f:daemonset-name":{},"f:pod-template-generation":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4be153d1-be77-4031-b20e-73937d0ff050\"}":{}}},"f:spec":{"f:affinity":{".":{},"f:nodeAffinity":{".":{},"f:requiredDuringSchedulingIgnoredDuringExecution":{}}},"f:containers":{"k:{\"name\":\"app\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:ports":{".":{},"k:{\"containerPort\":9376,\"protocol\":\"TCP\"}":{".":{},"f:containerPort":{},"f:protocol":{}}},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{},"f:tolerations":{}}}},{"manager":"kubelet","operation":"Update","apiVersion":"v1","time":"2022-12-14T16:21:56Z","fieldsType":"FieldsV1","fieldsV1":{"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.93\"}":{".":{},"f:ip":{}}},"f:startTime":{}}},"subresource":"status"}]},"spec":{"volumes":[{"name":"kube-api-access-95z2g","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"app","image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","ports":[{"containerPort":9376,"protocol":"TCP"}],"resources":{},"volumeMounts":[{"name":"kube-api-access-95z2g","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"File","imagePullPolicy":"IfNotPresent","securityContext":{}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","serviceAccountName":"default","serviceAccount":"default","nodeName":"kiedai4iakae-2","securityContext":{},"affinity":{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchFields":[{"key":"metadata.name","operator":"In","values":["kiedai4iakae-2"]}]}]}}},"schedulerName":"default-scheduler","tolerations":[{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/disk-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/pid-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unschedulable","operator":"Exists","effect":"NoSchedule"}],"priority":0,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:54Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:56Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:56Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2022-12-14T16:21:54Z"}],"hostIP":"192.168.121.197","podIP":"10.233.65.93","podIPs":[{"ip":"10.233.65.93"}],"startTime":"2022-12-14T16:21:54Z","containerStatuses":[{"name":"app","state":{"running":{"startedAt":"2022-12-14T16:21:55Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","imageID":"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3","containerID":"cri-o://94d323b9905a77ef322ce7e37b1c8f77e2d36bb68acca8f12b214b91ed17b4f3","started":true}],"qosClass":"BestEffort"}}]} - - [AfterEach] [sig-apps] Daemon set [Serial] + STEP: Creating a kubernetes client 02/25/23 10:19:34.97 + Feb 25 10:19:34.971: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 10:19:34.972 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:19:35.008 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:19:35.015 + [BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 + [It] should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance] + test/e2e/network/service.go:2157 + STEP: creating service in namespace services-542 02/25/23 10:19:35.022 + STEP: creating service affinity-clusterip in namespace services-542 02/25/23 10:19:35.022 + STEP: creating replication controller affinity-clusterip in namespace services-542 02/25/23 10:19:35.052 + I0225 10:19:35.069143 13 runners.go:193] Created replication controller with name: affinity-clusterip, namespace: services-542, replica count: 3 + I0225 10:19:38.121936 13 runners.go:193] affinity-clusterip Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + Feb 25 10:19:38.140: INFO: Creating new exec pod + Feb 25 10:19:38.160: INFO: Waiting up to 5m0s for pod "execpod-affinityxwzhq" in namespace "services-542" to be "running" + Feb 25 10:19:38.170: INFO: Pod "execpod-affinityxwzhq": Phase="Pending", Reason="", readiness=false. Elapsed: 9.091165ms + Feb 25 10:19:40.182: INFO: Pod "execpod-affinityxwzhq": Phase="Running", Reason="", readiness=true. Elapsed: 2.021654156s + Feb 25 10:19:40.182: INFO: Pod "execpod-affinityxwzhq" satisfied condition "running" + Feb 25 10:19:41.184: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-542 exec execpod-affinityxwzhq -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-clusterip 80' + Feb 25 10:19:41.589: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-clusterip 80\nConnection to affinity-clusterip 80 port [tcp/http] succeeded!\n" + Feb 25 10:19:41.589: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 10:19:41.589: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-542 exec execpod-affinityxwzhq -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.32.118 80' + Feb 25 10:19:42.073: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.32.118 80\nConnection to 10.233.32.118 80 port [tcp/http] succeeded!\n" + Feb 25 10:19:42.073: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 10:19:42.074: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-542 exec execpod-affinityxwzhq -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.32.118:80/ ; done' + Feb 25 10:19:42.808: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.32.118:80/\n" + Feb 25 10:19:42.808: INFO: stdout: "\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9\naffinity-clusterip-68vn9" + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Received response from host: affinity-clusterip-68vn9 + Feb 25 10:19:42.808: INFO: Cleaning up the exec pod + STEP: deleting ReplicationController affinity-clusterip in namespace services-542, will wait for the garbage collector to delete the pods 02/25/23 10:19:42.847 + Feb 25 10:19:42.927: INFO: Deleting ReplicationController affinity-clusterip took: 16.734723ms + Feb 25 10:19:43.029: INFO: Terminating ReplicationController affinity-clusterip pods took: 102.194232ms + [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 16:21:57.063: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "daemonsets-1968" for this suite. 12/14/22 16:21:57.08 + Feb 25 10:19:45.772: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-542" for this suite. 02/25/23 10:19:45.788 + [AfterEach] [sig-network] Services + test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services - should be able to change the type from NodePort to ExternalName [Conformance] - test/e2e/network/service.go:1523 + should be able to change the type from ExternalName to ClusterIP [Conformance] + test/e2e/network/service.go:1404 [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:21:57.097 -Dec 14 16:21:57.097: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 16:21:57.1 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:57.124 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:57.128 +STEP: Creating a kubernetes client 02/25/23 10:19:45.83 +Feb 25 10:19:45.831: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 10:19:45.837 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:19:45.883 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:19:45.889 [BeforeEach] [sig-network] Services test/e2e/network/service.go:758 -[It] should be able to change the type from NodePort to ExternalName [Conformance] - test/e2e/network/service.go:1523 -STEP: creating a service nodeport-service with the type=NodePort in namespace services-7345 12/14/22 16:21:57.133 -STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service 12/14/22 16:21:57.157 -STEP: creating service externalsvc in namespace services-7345 12/14/22 16:21:57.157 -STEP: creating replication controller externalsvc in namespace services-7345 12/14/22 16:21:57.178 -I1214 16:21:57.215453 14 runners.go:193] Created replication controller with name: externalsvc, namespace: services-7345, replica count: 2 -I1214 16:22:00.266371 14 runners.go:193] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -STEP: changing the NodePort service to type=ExternalName 12/14/22 16:22:00.273 -Dec 14 16:22:00.311: INFO: Creating new exec pod -Dec 14 16:22:00.322: INFO: Waiting up to 5m0s for pod "execpodcfqd5" in namespace "services-7345" to be "running" -Dec 14 16:22:00.328: INFO: Pod "execpodcfqd5": Phase="Pending", Reason="", readiness=false. Elapsed: 5.810762ms -Dec 14 16:22:02.334: INFO: Pod "execpodcfqd5": Phase="Running", Reason="", readiness=true. Elapsed: 2.011761898s -Dec 14 16:22:02.334: INFO: Pod "execpodcfqd5" satisfied condition "running" -Dec 14 16:22:02.335: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-7345 exec execpodcfqd5 -- /bin/sh -x -c nslookup nodeport-service.services-7345.svc.cluster.local' -Dec 14 16:22:02.751: INFO: stderr: "+ nslookup nodeport-service.services-7345.svc.cluster.local\n" -Dec 14 16:22:02.751: INFO: stdout: "Server:\t\t10.233.0.10\nAddress:\t10.233.0.10#53\n\nnodeport-service.services-7345.svc.cluster.local\tcanonical name = externalsvc.services-7345.svc.cluster.local.\nName:\texternalsvc.services-7345.svc.cluster.local\nAddress: 10.233.20.77\n\n" -STEP: deleting ReplicationController externalsvc in namespace services-7345, will wait for the garbage collector to delete the pods 12/14/22 16:22:02.751 -Dec 14 16:22:02.826: INFO: Deleting ReplicationController externalsvc took: 9.544298ms -Dec 14 16:22:02.927: INFO: Terminating ReplicationController externalsvc pods took: 100.764647ms -Dec 14 16:22:04.756: INFO: Cleaning up the NodePort to ExternalName test service +[It] should be able to change the type from ExternalName to ClusterIP [Conformance] + test/e2e/network/service.go:1404 +STEP: creating a service externalname-service with the type=ExternalName in namespace services-9373 02/25/23 10:19:45.897 +STEP: changing the ExternalName service to type=ClusterIP 02/25/23 10:19:45.91 +STEP: creating replication controller externalname-service in namespace services-9373 02/25/23 10:19:45.947 +I0225 10:19:45.968923 13 runners.go:193] Created replication controller with name: externalname-service, namespace: services-9373, replica count: 2 +I0225 10:19:49.020489 13 runners.go:193] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Feb 25 10:19:49.020: INFO: Creating new exec pod +Feb 25 10:19:49.061: INFO: Waiting up to 5m0s for pod "execpodnc44b" in namespace "services-9373" to be "running" +Feb 25 10:19:49.069: INFO: Pod "execpodnc44b": Phase="Pending", Reason="", readiness=false. Elapsed: 8.19515ms +Feb 25 10:19:51.092: INFO: Pod "execpodnc44b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030348148s +Feb 25 10:19:53.077: INFO: Pod "execpodnc44b": Phase="Running", Reason="", readiness=true. Elapsed: 4.016053846s +Feb 25 10:19:53.077: INFO: Pod "execpodnc44b" satisfied condition "running" +Feb 25 10:19:54.078: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9373 exec execpodnc44b -- /bin/sh -x -c echo hostName | nc -v -t -w 2 externalname-service 80' +Feb 25 10:19:54.396: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n" +Feb 25 10:19:54.396: INFO: stdout: "" +Feb 25 10:19:55.397: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9373 exec execpodnc44b -- /bin/sh -x -c echo hostName | nc -v -t -w 2 externalname-service 80' +Feb 25 10:19:55.830: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n" +Feb 25 10:19:55.830: INFO: stdout: "externalname-service-bsdpn" +Feb 25 10:19:55.831: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9373 exec execpodnc44b -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.42.235 80' +Feb 25 10:19:56.200: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.42.235 80\nConnection to 10.233.42.235 80 port [tcp/http] succeeded!\n" +Feb 25 10:19:56.200: INFO: stdout: "externalname-service-ctz94" +Feb 25 10:19:56.200: INFO: Cleaning up the ExternalName to ClusterIP test service [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 16:22:04.780: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-7345" for this suite. 12/14/22 16:22:04.788 +Feb 25 10:19:56.251: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-9373" for this suite. 02/25/23 10:19:56.267 [AfterEach] [sig-network] Services test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should be able to change the type from NodePort to ExternalName [Conformance]","completed":125,"skipped":2278,"failed":0} +{"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to ClusterIP [Conformance]","completed":118,"skipped":2173,"failed":0} ------------------------------ -• [SLOW TEST] [7.725 seconds] +• [SLOW TEST] [10.460 seconds] [sig-network] Services test/e2e/network/common/framework.go:23 - should be able to change the type from NodePort to ExternalName [Conformance] - test/e2e/network/service.go:1523 + should be able to change the type from ExternalName to ClusterIP [Conformance] + test/e2e/network/service.go:1404 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:21:57.097 - Dec 14 16:21:57.097: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 16:21:57.1 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:21:57.124 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:21:57.128 + STEP: Creating a kubernetes client 02/25/23 10:19:45.83 + Feb 25 10:19:45.831: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 10:19:45.837 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:19:45.883 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:19:45.889 [BeforeEach] [sig-network] Services test/e2e/network/service.go:758 - [It] should be able to change the type from NodePort to ExternalName [Conformance] - test/e2e/network/service.go:1523 - STEP: creating a service nodeport-service with the type=NodePort in namespace services-7345 12/14/22 16:21:57.133 - STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service 12/14/22 16:21:57.157 - STEP: creating service externalsvc in namespace services-7345 12/14/22 16:21:57.157 - STEP: creating replication controller externalsvc in namespace services-7345 12/14/22 16:21:57.178 - I1214 16:21:57.215453 14 runners.go:193] Created replication controller with name: externalsvc, namespace: services-7345, replica count: 2 - I1214 16:22:00.266371 14 runners.go:193] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - STEP: changing the NodePort service to type=ExternalName 12/14/22 16:22:00.273 - Dec 14 16:22:00.311: INFO: Creating new exec pod - Dec 14 16:22:00.322: INFO: Waiting up to 5m0s for pod "execpodcfqd5" in namespace "services-7345" to be "running" - Dec 14 16:22:00.328: INFO: Pod "execpodcfqd5": Phase="Pending", Reason="", readiness=false. Elapsed: 5.810762ms - Dec 14 16:22:02.334: INFO: Pod "execpodcfqd5": Phase="Running", Reason="", readiness=true. Elapsed: 2.011761898s - Dec 14 16:22:02.334: INFO: Pod "execpodcfqd5" satisfied condition "running" - Dec 14 16:22:02.335: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-7345 exec execpodcfqd5 -- /bin/sh -x -c nslookup nodeport-service.services-7345.svc.cluster.local' - Dec 14 16:22:02.751: INFO: stderr: "+ nslookup nodeport-service.services-7345.svc.cluster.local\n" - Dec 14 16:22:02.751: INFO: stdout: "Server:\t\t10.233.0.10\nAddress:\t10.233.0.10#53\n\nnodeport-service.services-7345.svc.cluster.local\tcanonical name = externalsvc.services-7345.svc.cluster.local.\nName:\texternalsvc.services-7345.svc.cluster.local\nAddress: 10.233.20.77\n\n" - STEP: deleting ReplicationController externalsvc in namespace services-7345, will wait for the garbage collector to delete the pods 12/14/22 16:22:02.751 - Dec 14 16:22:02.826: INFO: Deleting ReplicationController externalsvc took: 9.544298ms - Dec 14 16:22:02.927: INFO: Terminating ReplicationController externalsvc pods took: 100.764647ms - Dec 14 16:22:04.756: INFO: Cleaning up the NodePort to ExternalName test service + [It] should be able to change the type from ExternalName to ClusterIP [Conformance] + test/e2e/network/service.go:1404 + STEP: creating a service externalname-service with the type=ExternalName in namespace services-9373 02/25/23 10:19:45.897 + STEP: changing the ExternalName service to type=ClusterIP 02/25/23 10:19:45.91 + STEP: creating replication controller externalname-service in namespace services-9373 02/25/23 10:19:45.947 + I0225 10:19:45.968923 13 runners.go:193] Created replication controller with name: externalname-service, namespace: services-9373, replica count: 2 + I0225 10:19:49.020489 13 runners.go:193] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + Feb 25 10:19:49.020: INFO: Creating new exec pod + Feb 25 10:19:49.061: INFO: Waiting up to 5m0s for pod "execpodnc44b" in namespace "services-9373" to be "running" + Feb 25 10:19:49.069: INFO: Pod "execpodnc44b": Phase="Pending", Reason="", readiness=false. Elapsed: 8.19515ms + Feb 25 10:19:51.092: INFO: Pod "execpodnc44b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030348148s + Feb 25 10:19:53.077: INFO: Pod "execpodnc44b": Phase="Running", Reason="", readiness=true. Elapsed: 4.016053846s + Feb 25 10:19:53.077: INFO: Pod "execpodnc44b" satisfied condition "running" + Feb 25 10:19:54.078: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9373 exec execpodnc44b -- /bin/sh -x -c echo hostName | nc -v -t -w 2 externalname-service 80' + Feb 25 10:19:54.396: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n" + Feb 25 10:19:54.396: INFO: stdout: "" + Feb 25 10:19:55.397: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9373 exec execpodnc44b -- /bin/sh -x -c echo hostName | nc -v -t -w 2 externalname-service 80' + Feb 25 10:19:55.830: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n" + Feb 25 10:19:55.830: INFO: stdout: "externalname-service-bsdpn" + Feb 25 10:19:55.831: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9373 exec execpodnc44b -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.42.235 80' + Feb 25 10:19:56.200: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.42.235 80\nConnection to 10.233.42.235 80 port [tcp/http] succeeded!\n" + Feb 25 10:19:56.200: INFO: stdout: "externalname-service-ctz94" + Feb 25 10:19:56.200: INFO: Cleaning up the ExternalName to ClusterIP test service [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 16:22:04.780: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-7345" for this suite. 12/14/22 16:22:04.788 + Feb 25 10:19:56.251: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-9373" for this suite. 02/25/23 10:19:56.267 [AfterEach] [sig-network] Services test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSS +SS ------------------------------ -[sig-api-machinery] ResourceQuota - should create a ResourceQuota and capture the life of a secret. [Conformance] - test/e2e/apimachinery/resource_quota.go:150 -[BeforeEach] [sig-api-machinery] ResourceQuota +[sig-network] EndpointSlice + should create Endpoints and EndpointSlices for Pods matching a Service [Conformance] + test/e2e/network/endpointslice.go:204 +[BeforeEach] [sig-network] EndpointSlice test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:22:04.831 -Dec 14 16:22:04.831: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename resourcequota 12/14/22 16:22:04.834 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:04.859 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:04.863 -[It] should create a ResourceQuota and capture the life of a secret. [Conformance] - test/e2e/apimachinery/resource_quota.go:150 -STEP: Discovering how many secrets are in namespace by default 12/14/22 16:22:04.881 -STEP: Counting existing ResourceQuota 12/14/22 16:22:09.887 -STEP: Creating a ResourceQuota 12/14/22 16:22:14.896 -STEP: Ensuring resource quota status is calculated 12/14/22 16:22:14.906 -STEP: Creating a Secret 12/14/22 16:22:16.916 -STEP: Ensuring resource quota status captures secret creation 12/14/22 16:22:16.932 -STEP: Deleting a secret 12/14/22 16:22:18.942 -STEP: Ensuring resource quota status released usage 12/14/22 16:22:18.957 -[AfterEach] [sig-api-machinery] ResourceQuota +STEP: Creating a kubernetes client 02/25/23 10:19:56.29 +Feb 25 10:19:56.290: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename endpointslice 02/25/23 10:19:56.294 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:19:56.345 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:19:56.357 +[BeforeEach] [sig-network] EndpointSlice + test/e2e/network/endpointslice.go:51 +[It] should create Endpoints and EndpointSlices for Pods matching a Service [Conformance] + test/e2e/network/endpointslice.go:204 +STEP: referencing a single matching pod 02/25/23 10:20:01.588 +STEP: referencing matching pods with named port 02/25/23 10:20:06.608 +STEP: creating empty Endpoints and EndpointSlices for no matching Pods 02/25/23 10:20:11.629 +STEP: recreating EndpointSlices after they've been deleted 02/25/23 10:20:16.649 +Feb 25 10:20:16.698: INFO: EndpointSlice for Service endpointslice-593/example-named-port not found +[AfterEach] [sig-network] EndpointSlice test/e2e/framework/framework.go:187 -Dec 14 16:22:20.965: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-7094" for this suite. 12/14/22 16:22:20.974 -{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a secret. [Conformance]","completed":126,"skipped":2292,"failed":0} +Feb 25 10:20:26.719: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "endpointslice-593" for this suite. 02/25/23 10:20:26.729 +{"msg":"PASSED [sig-network] EndpointSlice should create Endpoints and EndpointSlices for Pods matching a Service [Conformance]","completed":119,"skipped":2175,"failed":0} ------------------------------ -• [SLOW TEST] [16.155 seconds] -[sig-api-machinery] ResourceQuota -test/e2e/apimachinery/framework.go:23 - should create a ResourceQuota and capture the life of a secret. [Conformance] - test/e2e/apimachinery/resource_quota.go:150 +• [SLOW TEST] [30.454 seconds] +[sig-network] EndpointSlice +test/e2e/network/common/framework.go:23 + should create Endpoints and EndpointSlices for Pods matching a Service [Conformance] + test/e2e/network/endpointslice.go:204 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] ResourceQuota + [BeforeEach] [sig-network] EndpointSlice test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:22:04.831 - Dec 14 16:22:04.831: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename resourcequota 12/14/22 16:22:04.834 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:04.859 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:04.863 - [It] should create a ResourceQuota and capture the life of a secret. [Conformance] - test/e2e/apimachinery/resource_quota.go:150 - STEP: Discovering how many secrets are in namespace by default 12/14/22 16:22:04.881 - STEP: Counting existing ResourceQuota 12/14/22 16:22:09.887 - STEP: Creating a ResourceQuota 12/14/22 16:22:14.896 - STEP: Ensuring resource quota status is calculated 12/14/22 16:22:14.906 - STEP: Creating a Secret 12/14/22 16:22:16.916 - STEP: Ensuring resource quota status captures secret creation 12/14/22 16:22:16.932 - STEP: Deleting a secret 12/14/22 16:22:18.942 - STEP: Ensuring resource quota status released usage 12/14/22 16:22:18.957 - [AfterEach] [sig-api-machinery] ResourceQuota + STEP: Creating a kubernetes client 02/25/23 10:19:56.29 + Feb 25 10:19:56.290: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename endpointslice 02/25/23 10:19:56.294 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:19:56.345 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:19:56.357 + [BeforeEach] [sig-network] EndpointSlice + test/e2e/network/endpointslice.go:51 + [It] should create Endpoints and EndpointSlices for Pods matching a Service [Conformance] + test/e2e/network/endpointslice.go:204 + STEP: referencing a single matching pod 02/25/23 10:20:01.588 + STEP: referencing matching pods with named port 02/25/23 10:20:06.608 + STEP: creating empty Endpoints and EndpointSlices for no matching Pods 02/25/23 10:20:11.629 + STEP: recreating EndpointSlices after they've been deleted 02/25/23 10:20:16.649 + Feb 25 10:20:16.698: INFO: EndpointSlice for Service endpointslice-593/example-named-port not found + [AfterEach] [sig-network] EndpointSlice test/e2e/framework/framework.go:187 - Dec 14 16:22:20.965: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "resourcequota-7094" for this suite. 12/14/22 16:22:20.974 + Feb 25 10:20:26.719: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "endpointslice-593" for this suite. 02/25/23 10:20:26.729 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSS +SS ------------------------------ -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - works for CRD preserving unknown fields at the schema root [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:193 -[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +[sig-apps] Job + should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] + test/e2e/apps/job.go:254 +[BeforeEach] [sig-apps] Job test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:22:20.992 -Dec 14 16:22:20.992: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 16:22:20.995 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:21.018 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:21.023 -[It] works for CRD preserving unknown fields at the schema root [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:193 -Dec 14 16:22:21.029: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: kubectl validation (kubectl create and apply) allows request with any unknown properties 12/14/22 16:22:24.327 -Dec 14 16:22:24.328: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2042 --namespace=crd-publish-openapi-2042 create -f -' -Dec 14 16:22:25.695: INFO: stderr: "" -Dec 14 16:22:25.695: INFO: stdout: "e2e-test-crd-publish-openapi-8103-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n" -Dec 14 16:22:25.695: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2042 --namespace=crd-publish-openapi-2042 delete e2e-test-crd-publish-openapi-8103-crds test-cr' -Dec 14 16:22:25.876: INFO: stderr: "" -Dec 14 16:22:25.876: INFO: stdout: "e2e-test-crd-publish-openapi-8103-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n" -Dec 14 16:22:25.877: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2042 --namespace=crd-publish-openapi-2042 apply -f -' -Dec 14 16:22:26.411: INFO: stderr: "" -Dec 14 16:22:26.411: INFO: stdout: "e2e-test-crd-publish-openapi-8103-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n" -Dec 14 16:22:26.411: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2042 --namespace=crd-publish-openapi-2042 delete e2e-test-crd-publish-openapi-8103-crds test-cr' -Dec 14 16:22:26.710: INFO: stderr: "" -Dec 14 16:22:26.710: INFO: stdout: "e2e-test-crd-publish-openapi-8103-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n" -STEP: kubectl explain works to explain CR 12/14/22 16:22:26.71 -Dec 14 16:22:26.710: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2042 explain e2e-test-crd-publish-openapi-8103-crds' -Dec 14 16:22:27.086: INFO: stderr: "" -Dec 14 16:22:27.086: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-8103-crd\nVERSION: crd-publish-openapi-test-unknown-at-root.example.com/v1\n\nDESCRIPTION:\n \n" -[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 10:20:26.745 +Feb 25 10:20:26.746: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename job 02/25/23 10:20:26.749 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:20:26.792 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:20:26.798 +[It] should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] + test/e2e/apps/job.go:254 +STEP: Creating a job 02/25/23 10:20:26.803 +STEP: Ensuring job reaches completions 02/25/23 10:20:26.816 +[AfterEach] [sig-apps] Job test/e2e/framework/framework.go:187 -Dec 14 16:22:30.051: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-publish-openapi-2042" for this suite. 12/14/22 16:22:30.068 -{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields at the schema root [Conformance]","completed":127,"skipped":2303,"failed":0} +Feb 25 10:20:40.834: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "job-3511" for this suite. 02/25/23 10:20:40.847 +{"msg":"PASSED [sig-apps] Job should run a job to completion when tasks sometimes fail and are locally restarted [Conformance]","completed":120,"skipped":2177,"failed":0} ------------------------------ -• [SLOW TEST] [9.094 seconds] -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +• [SLOW TEST] [14.115 seconds] +[sig-apps] Job +test/e2e/apps/framework.go:23 + should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] + test/e2e/apps/job.go:254 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-apps] Job + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:20:26.745 + Feb 25 10:20:26.746: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename job 02/25/23 10:20:26.749 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:20:26.792 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:20:26.798 + [It] should run a job to completion when tasks sometimes fail and are locally restarted [Conformance] + test/e2e/apps/job.go:254 + STEP: Creating a job 02/25/23 10:20:26.803 + STEP: Ensuring job reaches completions 02/25/23 10:20:26.816 + [AfterEach] [sig-apps] Job + test/e2e/framework/framework.go:187 + Feb 25 10:20:40.834: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "job-3511" for this suite. 02/25/23 10:20:40.847 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSSSS +------------------------------ +[sig-api-machinery] ResourceQuota + should create a ResourceQuota and capture the life of a replica set. [Conformance] + test/e2e/apimachinery/resource_quota.go:438 +[BeforeEach] [sig-api-machinery] ResourceQuota + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:20:40.878 +Feb 25 10:20:40.878: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename resourcequota 02/25/23 10:20:40.88 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:20:40.96 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:20:40.965 +[It] should create a ResourceQuota and capture the life of a replica set. [Conformance] + test/e2e/apimachinery/resource_quota.go:438 +STEP: Counting existing ResourceQuota 02/25/23 10:20:40.981 +STEP: Creating a ResourceQuota 02/25/23 10:20:45.988 +STEP: Ensuring resource quota status is calculated 02/25/23 10:20:46.001 +STEP: Creating a ReplicaSet 02/25/23 10:20:48.018 +STEP: Ensuring resource quota status captures replicaset creation 02/25/23 10:20:48.044 +STEP: Deleting a ReplicaSet 02/25/23 10:20:50.052 +STEP: Ensuring resource quota status released usage 02/25/23 10:20:50.077 +[AfterEach] [sig-api-machinery] ResourceQuota + test/e2e/framework/framework.go:187 +Feb 25 10:20:52.087: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-105" for this suite. 02/25/23 10:20:52.097 +{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replica set. [Conformance]","completed":121,"skipped":2194,"failed":0} +------------------------------ +• [SLOW TEST] [11.241 seconds] +[sig-api-machinery] ResourceQuota test/e2e/apimachinery/framework.go:23 - works for CRD preserving unknown fields at the schema root [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:193 + should create a ResourceQuota and capture the life of a replica set. [Conformance] + test/e2e/apimachinery/resource_quota.go:438 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + [BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:22:20.992 - Dec 14 16:22:20.992: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 16:22:20.995 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:21.018 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:21.023 - [It] works for CRD preserving unknown fields at the schema root [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:193 - Dec 14 16:22:21.029: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: kubectl validation (kubectl create and apply) allows request with any unknown properties 12/14/22 16:22:24.327 - Dec 14 16:22:24.328: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2042 --namespace=crd-publish-openapi-2042 create -f -' - Dec 14 16:22:25.695: INFO: stderr: "" - Dec 14 16:22:25.695: INFO: stdout: "e2e-test-crd-publish-openapi-8103-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n" - Dec 14 16:22:25.695: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2042 --namespace=crd-publish-openapi-2042 delete e2e-test-crd-publish-openapi-8103-crds test-cr' - Dec 14 16:22:25.876: INFO: stderr: "" - Dec 14 16:22:25.876: INFO: stdout: "e2e-test-crd-publish-openapi-8103-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n" - Dec 14 16:22:25.877: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2042 --namespace=crd-publish-openapi-2042 apply -f -' - Dec 14 16:22:26.411: INFO: stderr: "" - Dec 14 16:22:26.411: INFO: stdout: "e2e-test-crd-publish-openapi-8103-crd.crd-publish-openapi-test-unknown-at-root.example.com/test-cr created\n" - Dec 14 16:22:26.411: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2042 --namespace=crd-publish-openapi-2042 delete e2e-test-crd-publish-openapi-8103-crds test-cr' - Dec 14 16:22:26.710: INFO: stderr: "" - Dec 14 16:22:26.710: INFO: stdout: "e2e-test-crd-publish-openapi-8103-crd.crd-publish-openapi-test-unknown-at-root.example.com \"test-cr\" deleted\n" - STEP: kubectl explain works to explain CR 12/14/22 16:22:26.71 - Dec 14 16:22:26.710: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-2042 explain e2e-test-crd-publish-openapi-8103-crds' - Dec 14 16:22:27.086: INFO: stderr: "" - Dec 14 16:22:27.086: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-8103-crd\nVERSION: crd-publish-openapi-test-unknown-at-root.example.com/v1\n\nDESCRIPTION:\n \n" - [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 10:20:40.878 + Feb 25 10:20:40.878: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename resourcequota 02/25/23 10:20:40.88 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:20:40.96 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:20:40.965 + [It] should create a ResourceQuota and capture the life of a replica set. [Conformance] + test/e2e/apimachinery/resource_quota.go:438 + STEP: Counting existing ResourceQuota 02/25/23 10:20:40.981 + STEP: Creating a ResourceQuota 02/25/23 10:20:45.988 + STEP: Ensuring resource quota status is calculated 02/25/23 10:20:46.001 + STEP: Creating a ReplicaSet 02/25/23 10:20:48.018 + STEP: Ensuring resource quota status captures replicaset creation 02/25/23 10:20:48.044 + STEP: Deleting a ReplicaSet 02/25/23 10:20:50.052 + STEP: Ensuring resource quota status released usage 02/25/23 10:20:50.077 + [AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 - Dec 14 16:22:30.051: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "crd-publish-openapi-2042" for this suite. 12/14/22 16:22:30.068 + Feb 25 10:20:52.087: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "resourcequota-105" for this suite. 02/25/23 10:20:52.097 << End Captured GinkgoWriter Output ------------------------------ -SSSSS +SSS ------------------------------ -[sig-storage] Projected downwardAPI - should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:67 -[BeforeEach] [sig-storage] Projected downwardAPI +[sig-apps] Daemon set [Serial] + should update pod when spec was updated and update strategy is RollingUpdate [Conformance] + test/e2e/apps/daemon_set.go:373 +[BeforeEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:22:30.087 -Dec 14 16:22:30.088: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 16:22:30.096 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:30.122 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:30.126 -[BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 -[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:67 -STEP: Creating a pod to test downward API volume plugin 12/14/22 16:22:30.13 -Dec 14 16:22:30.147: INFO: Waiting up to 5m0s for pod "downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913" in namespace "projected-594" to be "Succeeded or Failed" -Dec 14 16:22:30.151: INFO: Pod "downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913": Phase="Pending", Reason="", readiness=false. Elapsed: 4.242244ms -Dec 14 16:22:32.157: INFO: Pod "downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01029205s -Dec 14 16:22:34.161: INFO: Pod "downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014266412s -STEP: Saw pod success 12/14/22 16:22:34.162 -Dec 14 16:22:34.162: INFO: Pod "downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913" satisfied condition "Succeeded or Failed" -Dec 14 16:22:34.167: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913 container client-container: -STEP: delete the pod 12/14/22 16:22:34.192 -Dec 14 16:22:34.224: INFO: Waiting for pod downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913 to disappear -Dec 14 16:22:34.230: INFO: Pod downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913 no longer exists -[AfterEach] [sig-storage] Projected downwardAPI +STEP: Creating a kubernetes client 02/25/23 10:20:52.123 +Feb 25 10:20:52.123: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename daemonsets 02/25/23 10:20:52.126 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:20:52.156 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:20:52.161 +[BeforeEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:145 +[It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] + test/e2e/apps/daemon_set.go:373 +Feb 25 10:20:52.210: INFO: Creating simple daemon set daemon-set +STEP: Check that daemon pods launch on every node of the cluster. 02/25/23 10:20:52.22 +Feb 25 10:20:52.235: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:20:52.235: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 10:20:53.316: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:20:53.316: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 10:20:54.257: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 10:20:54.257: INFO: Node diek3dua9gay-2 is running 0 daemon pod, expected 1 +Feb 25 10:20:55.252: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 +Feb 25 10:20:55.252: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set +STEP: Update daemon pods image. 02/25/23 10:20:55.281 +STEP: Check that daemon pods images are updated. 02/25/23 10:20:55.301 +Feb 25 10:20:55.309: INFO: Wrong image for pod: daemon-set-6gzqk. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. +Feb 25 10:20:55.309: INFO: Wrong image for pod: daemon-set-l897d. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. +Feb 25 10:20:55.309: INFO: Wrong image for pod: daemon-set-mxgpb. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. +Feb 25 10:20:56.326: INFO: Wrong image for pod: daemon-set-l897d. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. +Feb 25 10:20:56.326: INFO: Wrong image for pod: daemon-set-mxgpb. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. +Feb 25 10:20:57.330: INFO: Wrong image for pod: daemon-set-l897d. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. +Feb 25 10:20:57.330: INFO: Pod daemon-set-mlvq6 is not available +Feb 25 10:20:57.330: INFO: Wrong image for pod: daemon-set-mxgpb. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. +Feb 25 10:20:58.333: INFO: Wrong image for pod: daemon-set-l897d. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. +Feb 25 10:20:58.333: INFO: Pod daemon-set-mlvq6 is not available +Feb 25 10:20:58.333: INFO: Wrong image for pod: daemon-set-mxgpb. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. +Feb 25 10:20:59.327: INFO: Wrong image for pod: daemon-set-mxgpb. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. +Feb 25 10:21:00.326: INFO: Wrong image for pod: daemon-set-mxgpb. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. +Feb 25 10:21:00.326: INFO: Pod daemon-set-v2vpk is not available +Feb 25 10:21:01.336: INFO: Wrong image for pod: daemon-set-mxgpb. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. +Feb 25 10:21:01.336: INFO: Pod daemon-set-v2vpk is not available +Feb 25 10:21:03.330: INFO: Pod daemon-set-9w77n is not available +STEP: Check that daemon pods are still running on every node of the cluster. 02/25/23 10:21:03.343 +Feb 25 10:21:03.364: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 10:21:03.364: INFO: Node diek3dua9gay-2 is running 0 daemon pod, expected 1 +Feb 25 10:21:04.385: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 10:21:04.385: INFO: Node diek3dua9gay-2 is running 0 daemon pod, expected 1 +Feb 25 10:21:05.384: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 +Feb 25 10:21:05.384: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set +[AfterEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:110 +STEP: Deleting DaemonSet "daemon-set" 02/25/23 10:21:05.44 +STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6053, will wait for the garbage collector to delete the pods 02/25/23 10:21:05.441 +Feb 25 10:21:05.522: INFO: Deleting DaemonSet.extensions daemon-set took: 22.507328ms +Feb 25 10:21:05.643: INFO: Terminating DaemonSet.extensions daemon-set pods took: 120.925537ms +Feb 25 10:21:07.654: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:21:07.655: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set +Feb 25 10:21:07.669: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"19288"},"items":null} + +Feb 25 10:21:07.675: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"19288"},"items":null} + +[AfterEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:187 -Dec 14 16:22:34.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-594" for this suite. 12/14/22 16:22:34.236 -{"msg":"PASSED [sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","completed":128,"skipped":2308,"failed":0} +Feb 25 10:21:07.725: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "daemonsets-6053" for this suite. 02/25/23 10:21:07.735 +{"msg":"PASSED [sig-apps] Daemon set [Serial] should update pod when spec was updated and update strategy is RollingUpdate [Conformance]","completed":122,"skipped":2197,"failed":0} ------------------------------ -• [4.156 seconds] -[sig-storage] Projected downwardAPI -test/e2e/common/storage/framework.go:23 - should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:67 +• [SLOW TEST] [15.625 seconds] +[sig-apps] Daemon set [Serial] +test/e2e/apps/framework.go:23 + should update pod when spec was updated and update strategy is RollingUpdate [Conformance] + test/e2e/apps/daemon_set.go:373 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected downwardAPI + [BeforeEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:22:30.087 - Dec 14 16:22:30.088: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 16:22:30.096 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:30.122 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:30.126 - [BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 - [It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:67 - STEP: Creating a pod to test downward API volume plugin 12/14/22 16:22:30.13 - Dec 14 16:22:30.147: INFO: Waiting up to 5m0s for pod "downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913" in namespace "projected-594" to be "Succeeded or Failed" - Dec 14 16:22:30.151: INFO: Pod "downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913": Phase="Pending", Reason="", readiness=false. Elapsed: 4.242244ms - Dec 14 16:22:32.157: INFO: Pod "downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01029205s - Dec 14 16:22:34.161: INFO: Pod "downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014266412s - STEP: Saw pod success 12/14/22 16:22:34.162 - Dec 14 16:22:34.162: INFO: Pod "downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913" satisfied condition "Succeeded or Failed" - Dec 14 16:22:34.167: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913 container client-container: - STEP: delete the pod 12/14/22 16:22:34.192 - Dec 14 16:22:34.224: INFO: Waiting for pod downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913 to disappear - Dec 14 16:22:34.230: INFO: Pod downwardapi-volume-a54460d6-3fb8-4c95-929d-c22927a8c913 no longer exists - [AfterEach] [sig-storage] Projected downwardAPI + STEP: Creating a kubernetes client 02/25/23 10:20:52.123 + Feb 25 10:20:52.123: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename daemonsets 02/25/23 10:20:52.126 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:20:52.156 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:20:52.161 + [BeforeEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:145 + [It] should update pod when spec was updated and update strategy is RollingUpdate [Conformance] + test/e2e/apps/daemon_set.go:373 + Feb 25 10:20:52.210: INFO: Creating simple daemon set daemon-set + STEP: Check that daemon pods launch on every node of the cluster. 02/25/23 10:20:52.22 + Feb 25 10:20:52.235: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:20:52.235: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 10:20:53.316: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:20:53.316: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 10:20:54.257: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 10:20:54.257: INFO: Node diek3dua9gay-2 is running 0 daemon pod, expected 1 + Feb 25 10:20:55.252: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 + Feb 25 10:20:55.252: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set + STEP: Update daemon pods image. 02/25/23 10:20:55.281 + STEP: Check that daemon pods images are updated. 02/25/23 10:20:55.301 + Feb 25 10:20:55.309: INFO: Wrong image for pod: daemon-set-6gzqk. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. + Feb 25 10:20:55.309: INFO: Wrong image for pod: daemon-set-l897d. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. + Feb 25 10:20:55.309: INFO: Wrong image for pod: daemon-set-mxgpb. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. + Feb 25 10:20:56.326: INFO: Wrong image for pod: daemon-set-l897d. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. + Feb 25 10:20:56.326: INFO: Wrong image for pod: daemon-set-mxgpb. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. + Feb 25 10:20:57.330: INFO: Wrong image for pod: daemon-set-l897d. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. + Feb 25 10:20:57.330: INFO: Pod daemon-set-mlvq6 is not available + Feb 25 10:20:57.330: INFO: Wrong image for pod: daemon-set-mxgpb. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. + Feb 25 10:20:58.333: INFO: Wrong image for pod: daemon-set-l897d. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. + Feb 25 10:20:58.333: INFO: Pod daemon-set-mlvq6 is not available + Feb 25 10:20:58.333: INFO: Wrong image for pod: daemon-set-mxgpb. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. + Feb 25 10:20:59.327: INFO: Wrong image for pod: daemon-set-mxgpb. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. + Feb 25 10:21:00.326: INFO: Wrong image for pod: daemon-set-mxgpb. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. + Feb 25 10:21:00.326: INFO: Pod daemon-set-v2vpk is not available + Feb 25 10:21:01.336: INFO: Wrong image for pod: daemon-set-mxgpb. Expected: registry.k8s.io/e2e-test-images/agnhost:2.40, got: registry.k8s.io/e2e-test-images/httpd:2.4.38-2. + Feb 25 10:21:01.336: INFO: Pod daemon-set-v2vpk is not available + Feb 25 10:21:03.330: INFO: Pod daemon-set-9w77n is not available + STEP: Check that daemon pods are still running on every node of the cluster. 02/25/23 10:21:03.343 + Feb 25 10:21:03.364: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 10:21:03.364: INFO: Node diek3dua9gay-2 is running 0 daemon pod, expected 1 + Feb 25 10:21:04.385: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 10:21:04.385: INFO: Node diek3dua9gay-2 is running 0 daemon pod, expected 1 + Feb 25 10:21:05.384: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 + Feb 25 10:21:05.384: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set + [AfterEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:110 + STEP: Deleting DaemonSet "daemon-set" 02/25/23 10:21:05.44 + STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6053, will wait for the garbage collector to delete the pods 02/25/23 10:21:05.441 + Feb 25 10:21:05.522: INFO: Deleting DaemonSet.extensions daemon-set took: 22.507328ms + Feb 25 10:21:05.643: INFO: Terminating DaemonSet.extensions daemon-set pods took: 120.925537ms + Feb 25 10:21:07.654: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:21:07.655: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set + Feb 25 10:21:07.669: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"19288"},"items":null} + + Feb 25 10:21:07.675: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"19288"},"items":null} + + [AfterEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:187 - Dec 14 16:22:34.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-594" for this suite. 12/14/22 16:22:34.236 + Feb 25 10:21:07.725: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "daemonsets-6053" for this suite. 02/25/23 10:21:07.735 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSSSSSSSS ------------------------------ -[sig-node] Variable Expansion - should fail substituting values in a volume subpath with backticks [Slow] [Conformance] - test/e2e/common/node/expansion.go:151 -[BeforeEach] [sig-node] Variable Expansion +[sig-node] Security Context + should support pod.Spec.SecurityContext.RunAsUser And pod.Spec.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] + test/e2e/node/security_context.go:97 +[BeforeEach] [sig-node] Security Context test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:22:34.247 -Dec 14 16:22:34.247: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename var-expansion 12/14/22 16:22:34.249 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:34.271 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:34.274 -[It] should fail substituting values in a volume subpath with backticks [Slow] [Conformance] - test/e2e/common/node/expansion.go:151 -Dec 14 16:22:34.291: INFO: Waiting up to 2m0s for pod "var-expansion-7c906a15-165a-4e32-b465-79d7f9aa6150" in namespace "var-expansion-621" to be "container 0 failed with reason CreateContainerConfigError" -Dec 14 16:22:34.295: INFO: Pod "var-expansion-7c906a15-165a-4e32-b465-79d7f9aa6150": Phase="Pending", Reason="", readiness=false. Elapsed: 3.678031ms -Dec 14 16:22:36.301: INFO: Pod "var-expansion-7c906a15-165a-4e32-b465-79d7f9aa6150": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010083904s -Dec 14 16:22:36.301: INFO: Pod "var-expansion-7c906a15-165a-4e32-b465-79d7f9aa6150" satisfied condition "container 0 failed with reason CreateContainerConfigError" -Dec 14 16:22:36.302: INFO: Deleting pod "var-expansion-7c906a15-165a-4e32-b465-79d7f9aa6150" in namespace "var-expansion-621" -Dec 14 16:22:36.313: INFO: Wait up to 5m0s for pod "var-expansion-7c906a15-165a-4e32-b465-79d7f9aa6150" to be fully deleted -[AfterEach] [sig-node] Variable Expansion +STEP: Creating a kubernetes client 02/25/23 10:21:07.755 +Feb 25 10:21:07.759: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename security-context 02/25/23 10:21:07.763 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:21:07.798 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:21:07.804 +[It] should support pod.Spec.SecurityContext.RunAsUser And pod.Spec.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] + test/e2e/node/security_context.go:97 +STEP: Creating a pod to test pod.Spec.SecurityContext.RunAsUser 02/25/23 10:21:07.81 +Feb 25 10:21:07.831: INFO: Waiting up to 5m0s for pod "security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7" in namespace "security-context-3748" to be "Succeeded or Failed" +Feb 25 10:21:07.846: INFO: Pod "security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7": Phase="Pending", Reason="", readiness=false. Elapsed: 14.645558ms +Feb 25 10:21:09.854: INFO: Pod "security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023186711s +Feb 25 10:21:11.854: INFO: Pod "security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022845528s +STEP: Saw pod success 02/25/23 10:21:11.854 +Feb 25 10:21:11.855: INFO: Pod "security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7" satisfied condition "Succeeded or Failed" +Feb 25 10:21:11.862: INFO: Trying to get logs from node diek3dua9gay-3 pod security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7 container test-container: +STEP: delete the pod 02/25/23 10:21:11.89 +Feb 25 10:21:11.916: INFO: Waiting for pod security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7 to disappear +Feb 25 10:21:11.925: INFO: Pod security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7 no longer exists +[AfterEach] [sig-node] Security Context test/e2e/framework/framework.go:187 -Dec 14 16:22:38.324: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "var-expansion-621" for this suite. 12/14/22 16:22:38.33 -{"msg":"PASSED [sig-node] Variable Expansion should fail substituting values in a volume subpath with backticks [Slow] [Conformance]","completed":129,"skipped":2311,"failed":0} +Feb 25 10:21:11.926: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "security-context-3748" for this suite. 02/25/23 10:21:11.939 +{"msg":"PASSED [sig-node] Security Context should support pod.Spec.SecurityContext.RunAsUser And pod.Spec.SecurityContext.RunAsGroup [LinuxOnly] [Conformance]","completed":123,"skipped":2209,"failed":0} ------------------------------ -• [4.092 seconds] -[sig-node] Variable Expansion -test/e2e/common/node/framework.go:23 - should fail substituting values in a volume subpath with backticks [Slow] [Conformance] - test/e2e/common/node/expansion.go:151 +• [4.197 seconds] +[sig-node] Security Context +test/e2e/node/framework.go:23 + should support pod.Spec.SecurityContext.RunAsUser And pod.Spec.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] + test/e2e/node/security_context.go:97 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Variable Expansion + [BeforeEach] [sig-node] Security Context test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:22:34.247 - Dec 14 16:22:34.247: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename var-expansion 12/14/22 16:22:34.249 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:34.271 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:34.274 - [It] should fail substituting values in a volume subpath with backticks [Slow] [Conformance] - test/e2e/common/node/expansion.go:151 - Dec 14 16:22:34.291: INFO: Waiting up to 2m0s for pod "var-expansion-7c906a15-165a-4e32-b465-79d7f9aa6150" in namespace "var-expansion-621" to be "container 0 failed with reason CreateContainerConfigError" - Dec 14 16:22:34.295: INFO: Pod "var-expansion-7c906a15-165a-4e32-b465-79d7f9aa6150": Phase="Pending", Reason="", readiness=false. Elapsed: 3.678031ms - Dec 14 16:22:36.301: INFO: Pod "var-expansion-7c906a15-165a-4e32-b465-79d7f9aa6150": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010083904s - Dec 14 16:22:36.301: INFO: Pod "var-expansion-7c906a15-165a-4e32-b465-79d7f9aa6150" satisfied condition "container 0 failed with reason CreateContainerConfigError" - Dec 14 16:22:36.302: INFO: Deleting pod "var-expansion-7c906a15-165a-4e32-b465-79d7f9aa6150" in namespace "var-expansion-621" - Dec 14 16:22:36.313: INFO: Wait up to 5m0s for pod "var-expansion-7c906a15-165a-4e32-b465-79d7f9aa6150" to be fully deleted - [AfterEach] [sig-node] Variable Expansion + STEP: Creating a kubernetes client 02/25/23 10:21:07.755 + Feb 25 10:21:07.759: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename security-context 02/25/23 10:21:07.763 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:21:07.798 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:21:07.804 + [It] should support pod.Spec.SecurityContext.RunAsUser And pod.Spec.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] + test/e2e/node/security_context.go:97 + STEP: Creating a pod to test pod.Spec.SecurityContext.RunAsUser 02/25/23 10:21:07.81 + Feb 25 10:21:07.831: INFO: Waiting up to 5m0s for pod "security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7" in namespace "security-context-3748" to be "Succeeded or Failed" + Feb 25 10:21:07.846: INFO: Pod "security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7": Phase="Pending", Reason="", readiness=false. Elapsed: 14.645558ms + Feb 25 10:21:09.854: INFO: Pod "security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023186711s + Feb 25 10:21:11.854: INFO: Pod "security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022845528s + STEP: Saw pod success 02/25/23 10:21:11.854 + Feb 25 10:21:11.855: INFO: Pod "security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7" satisfied condition "Succeeded or Failed" + Feb 25 10:21:11.862: INFO: Trying to get logs from node diek3dua9gay-3 pod security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7 container test-container: + STEP: delete the pod 02/25/23 10:21:11.89 + Feb 25 10:21:11.916: INFO: Waiting for pod security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7 to disappear + Feb 25 10:21:11.925: INFO: Pod security-context-5a4cdc5d-21d1-4171-9a7b-cfa5a82bf3f7 no longer exists + [AfterEach] [sig-node] Security Context test/e2e/framework/framework.go:187 - Dec 14 16:22:38.324: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "var-expansion-621" for this suite. 12/14/22 16:22:38.33 + Feb 25 10:21:11.926: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "security-context-3748" for this suite. 02/25/23 10:21:11.939 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] ResourceQuota - should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] - test/e2e/apimachinery/resource_quota.go:65 -[BeforeEach] [sig-api-machinery] ResourceQuota +[sig-storage] Subpath Atomic writer volumes + should support subpaths with configmap pod [Conformance] + test/e2e/storage/subpath.go:70 +[BeforeEach] [sig-storage] Subpath test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:22:38.348 -Dec 14 16:22:38.348: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename resourcequota 12/14/22 16:22:38.35 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:38.375 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:38.378 -[It] should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] - test/e2e/apimachinery/resource_quota.go:65 -STEP: Counting existing ResourceQuota 12/14/22 16:22:38.382 -STEP: Creating a ResourceQuota 12/14/22 16:22:43.391 -STEP: Ensuring resource quota status is calculated 12/14/22 16:22:43.403 -[AfterEach] [sig-api-machinery] ResourceQuota +STEP: Creating a kubernetes client 02/25/23 10:21:11.975 +Feb 25 10:21:11.975: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename subpath 02/25/23 10:21:11.98 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:21:12.021 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:21:12.027 +[BeforeEach] Atomic writer volumes + test/e2e/storage/subpath.go:40 +STEP: Setting up data 02/25/23 10:21:12.031 +[It] should support subpaths with configmap pod [Conformance] + test/e2e/storage/subpath.go:70 +STEP: Creating pod pod-subpath-test-configmap-lfk9 02/25/23 10:21:12.061 +STEP: Creating a pod to test atomic-volume-subpath 02/25/23 10:21:12.061 +Feb 25 10:21:12.080: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-lfk9" in namespace "subpath-4976" to be "Succeeded or Failed" +Feb 25 10:21:12.086: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Pending", Reason="", readiness=false. Elapsed: 6.747034ms +Feb 25 10:21:14.094: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 2.01485851s +Feb 25 10:21:16.095: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 4.015410582s +Feb 25 10:21:18.096: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 6.016548244s +Feb 25 10:21:20.099: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 8.01970693s +Feb 25 10:21:22.094: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 10.01472838s +Feb 25 10:21:24.098: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 12.01879177s +Feb 25 10:21:26.095: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 14.015821089s +Feb 25 10:21:28.096: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 16.016250325s +Feb 25 10:21:30.097: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 18.017333873s +Feb 25 10:21:32.099: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 20.01959294s +Feb 25 10:21:34.096: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=false. Elapsed: 22.016209558s +Feb 25 10:21:36.095: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.015263485s +STEP: Saw pod success 02/25/23 10:21:36.095 +Feb 25 10:21:36.095: INFO: Pod "pod-subpath-test-configmap-lfk9" satisfied condition "Succeeded or Failed" +Feb 25 10:21:36.105: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-subpath-test-configmap-lfk9 container test-container-subpath-configmap-lfk9: +STEP: delete the pod 02/25/23 10:21:36.128 +Feb 25 10:21:36.155: INFO: Waiting for pod pod-subpath-test-configmap-lfk9 to disappear +Feb 25 10:21:36.165: INFO: Pod pod-subpath-test-configmap-lfk9 no longer exists +STEP: Deleting pod pod-subpath-test-configmap-lfk9 02/25/23 10:21:36.165 +Feb 25 10:21:36.166: INFO: Deleting pod "pod-subpath-test-configmap-lfk9" in namespace "subpath-4976" +[AfterEach] [sig-storage] Subpath test/e2e/framework/framework.go:187 -Dec 14 16:22:45.416: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-3078" for this suite. 12/14/22 16:22:45.424 -{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]","completed":130,"skipped":2342,"failed":0} +Feb 25 10:21:36.175: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "subpath-4976" for this suite. 02/25/23 10:21:36.193 +{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod [Conformance]","completed":124,"skipped":2299,"failed":0} ------------------------------ -• [SLOW TEST] [7.086 seconds] -[sig-api-machinery] ResourceQuota -test/e2e/apimachinery/framework.go:23 - should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] - test/e2e/apimachinery/resource_quota.go:65 +• [SLOW TEST] [24.236 seconds] +[sig-storage] Subpath +test/e2e/storage/utils/framework.go:23 + Atomic writer volumes + test/e2e/storage/subpath.go:36 + should support subpaths with configmap pod [Conformance] + test/e2e/storage/subpath.go:70 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] ResourceQuota + [BeforeEach] [sig-storage] Subpath test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:22:38.348 - Dec 14 16:22:38.348: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename resourcequota 12/14/22 16:22:38.35 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:38.375 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:38.378 - [It] should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] - test/e2e/apimachinery/resource_quota.go:65 - STEP: Counting existing ResourceQuota 12/14/22 16:22:38.382 - STEP: Creating a ResourceQuota 12/14/22 16:22:43.391 - STEP: Ensuring resource quota status is calculated 12/14/22 16:22:43.403 - [AfterEach] [sig-api-machinery] ResourceQuota + STEP: Creating a kubernetes client 02/25/23 10:21:11.975 + Feb 25 10:21:11.975: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename subpath 02/25/23 10:21:11.98 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:21:12.021 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:21:12.027 + [BeforeEach] Atomic writer volumes + test/e2e/storage/subpath.go:40 + STEP: Setting up data 02/25/23 10:21:12.031 + [It] should support subpaths with configmap pod [Conformance] + test/e2e/storage/subpath.go:70 + STEP: Creating pod pod-subpath-test-configmap-lfk9 02/25/23 10:21:12.061 + STEP: Creating a pod to test atomic-volume-subpath 02/25/23 10:21:12.061 + Feb 25 10:21:12.080: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-lfk9" in namespace "subpath-4976" to be "Succeeded or Failed" + Feb 25 10:21:12.086: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Pending", Reason="", readiness=false. Elapsed: 6.747034ms + Feb 25 10:21:14.094: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 2.01485851s + Feb 25 10:21:16.095: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 4.015410582s + Feb 25 10:21:18.096: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 6.016548244s + Feb 25 10:21:20.099: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 8.01970693s + Feb 25 10:21:22.094: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 10.01472838s + Feb 25 10:21:24.098: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 12.01879177s + Feb 25 10:21:26.095: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 14.015821089s + Feb 25 10:21:28.096: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 16.016250325s + Feb 25 10:21:30.097: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 18.017333873s + Feb 25 10:21:32.099: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=true. Elapsed: 20.01959294s + Feb 25 10:21:34.096: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Running", Reason="", readiness=false. Elapsed: 22.016209558s + Feb 25 10:21:36.095: INFO: Pod "pod-subpath-test-configmap-lfk9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.015263485s + STEP: Saw pod success 02/25/23 10:21:36.095 + Feb 25 10:21:36.095: INFO: Pod "pod-subpath-test-configmap-lfk9" satisfied condition "Succeeded or Failed" + Feb 25 10:21:36.105: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-subpath-test-configmap-lfk9 container test-container-subpath-configmap-lfk9: + STEP: delete the pod 02/25/23 10:21:36.128 + Feb 25 10:21:36.155: INFO: Waiting for pod pod-subpath-test-configmap-lfk9 to disappear + Feb 25 10:21:36.165: INFO: Pod pod-subpath-test-configmap-lfk9 no longer exists + STEP: Deleting pod pod-subpath-test-configmap-lfk9 02/25/23 10:21:36.165 + Feb 25 10:21:36.166: INFO: Deleting pod "pod-subpath-test-configmap-lfk9" in namespace "subpath-4976" + [AfterEach] [sig-storage] Subpath test/e2e/framework/framework.go:187 - Dec 14 16:22:45.416: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "resourcequota-3078" for this suite. 12/14/22 16:22:45.424 + Feb 25 10:21:36.175: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "subpath-4976" for this suite. 02/25/23 10:21:36.193 << End Captured GinkgoWriter Output ------------------------------ -SSSSSS +SSSSSSSSSSSSSSSS ------------------------------ -[sig-network] Proxy version v1 - should proxy through a service and a pod [Conformance] - test/e2e/network/proxy.go:101 -[BeforeEach] version v1 +[sig-network] DNS + should provide DNS for pods for Hostname [Conformance] + test/e2e/network/dns.go:248 +[BeforeEach] [sig-network] DNS test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:22:45.434 -Dec 14 16:22:45.435: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename proxy 12/14/22 16:22:45.436 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:45.464 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:45.467 -[It] should proxy through a service and a pod [Conformance] - test/e2e/network/proxy.go:101 -STEP: starting an echo server on multiple ports 12/14/22 16:22:45.487 -STEP: creating replication controller proxy-service-xn49c in namespace proxy-6092 12/14/22 16:22:45.487 -I1214 16:22:45.499732 14 runners.go:193] Created replication controller with name: proxy-service-xn49c, namespace: proxy-6092, replica count: 1 -I1214 16:22:46.550851 14 runners.go:193] proxy-service-xn49c Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -I1214 16:22:47.551380 14 runners.go:193] proxy-service-xn49c Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady -I1214 16:22:48.552422 14 runners.go:193] proxy-service-xn49c Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -Dec 14 16:22:48.560: INFO: setup took 3.088408066s, starting test cases -STEP: running 16 cases, 20 attempts per case, 320 total attempts 12/14/22 16:22:48.56 -Dec 14 16:22:48.574: INFO: (0) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 13.621855ms) -Dec 14 16:22:48.583: INFO: (0) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 23.122439ms) -Dec 14 16:22:48.584: INFO: (0) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 22.864897ms) -Dec 14 16:22:48.584: INFO: (0) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 23.402371ms) -Dec 14 16:22:48.584: INFO: (0) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 22.538269ms) -Dec 14 16:22:48.584: INFO: (0) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 22.516574ms) -Dec 14 16:22:48.584: INFO: (0) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 22.503565ms) -Dec 14 16:22:48.584: INFO: (0) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 22.766861ms) -Dec 14 16:22:48.589: INFO: (0) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 28.430669ms) -Dec 14 16:22:48.589: INFO: (0) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 27.387612ms) -Dec 14 16:22:48.593: INFO: (0) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 31.935466ms) -Dec 14 16:22:48.593: INFO: (0) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 32.474889ms) -Dec 14 16:22:48.593: INFO: (0) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test (200; 12.608289ms) -Dec 14 16:22:48.609: INFO: (1) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 12.283916ms) -Dec 14 16:22:48.609: INFO: (1) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 13.385954ms) -Dec 14 16:22:48.610: INFO: (1) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 13.829023ms) -Dec 14 16:22:48.612: INFO: (1) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 15.284702ms) -Dec 14 16:22:48.613: INFO: (1) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 16.63793ms) -Dec 14 16:22:48.615: INFO: (1) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 18.094382ms) -Dec 14 16:22:48.615: INFO: (1) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 18.700518ms) -Dec 14 16:22:48.615: INFO: (1) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: ... (200; 17.981849ms) -Dec 14 16:22:48.644: INFO: (2) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 19.011543ms) -Dec 14 16:22:48.644: INFO: (2) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 21.556156ms) -Dec 14 16:22:48.648: INFO: (2) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 23.277576ms) -Dec 14 16:22:48.648: INFO: (2) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 23.317741ms) -Dec 14 16:22:48.649: INFO: (2) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 23.675129ms) -Dec 14 16:22:48.649: INFO: (2) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 24.270845ms) -Dec 14 16:22:48.649: INFO: (2) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 24.440642ms) -Dec 14 16:22:48.649: INFO: (2) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 24.135461ms) -Dec 14 16:22:48.649: INFO: (2) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 24.535369ms) -Dec 14 16:22:48.649: INFO: (2) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 23.644196ms) -Dec 14 16:22:48.654: INFO: (2) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 29.532621ms) -Dec 14 16:22:48.655: INFO: (2) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 29.528768ms) -Dec 14 16:22:48.655: INFO: (2) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 30.537116ms) -Dec 14 16:22:48.667: INFO: (3) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 10.888789ms) -Dec 14 16:22:48.668: INFO: (3) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test (200; 14.12124ms) -Dec 14 16:22:48.670: INFO: (3) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 14.094639ms) -Dec 14 16:22:48.670: INFO: (3) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 13.799731ms) -Dec 14 16:22:48.670: INFO: (3) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 13.914116ms) -Dec 14 16:22:48.670: INFO: (3) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 13.783736ms) -Dec 14 16:22:48.671: INFO: (3) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 14.43458ms) -Dec 14 16:22:48.673: INFO: (3) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 17.180541ms) -Dec 14 16:22:48.674: INFO: (3) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 17.249456ms) -Dec 14 16:22:48.675: INFO: (3) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 18.941391ms) -Dec 14 16:22:48.675: INFO: (3) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 18.921959ms) -Dec 14 16:22:48.675: INFO: (3) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 19.103763ms) -Dec 14 16:22:48.676: INFO: (3) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 18.979113ms) -Dec 14 16:22:48.676: INFO: (3) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 19.383251ms) -Dec 14 16:22:48.679: INFO: (3) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 22.379524ms) -Dec 14 16:22:48.692: INFO: (4) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 12.535147ms) -Dec 14 16:22:48.693: INFO: (4) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 13.233753ms) -Dec 14 16:22:48.694: INFO: (4) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 14.63415ms) -Dec 14 16:22:48.694: INFO: (4) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 14.361648ms) -Dec 14 16:22:48.695: INFO: (4) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 14.112028ms) -Dec 14 16:22:48.701: INFO: (4) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 22.162349ms) -Dec 14 16:22:48.701: INFO: (4) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 20.989005ms) -Dec 14 16:22:48.701: INFO: (4) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 21.970152ms) -Dec 14 16:22:48.702: INFO: (4) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 23.214586ms) -Dec 14 16:22:48.702: INFO: (4) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 22.425565ms) -Dec 14 16:22:48.702: INFO: (4) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 16.668387ms) -Dec 14 16:22:48.720: INFO: (5) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 16.48941ms) -Dec 14 16:22:48.720: INFO: (5) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 16.960944ms) -Dec 14 16:22:48.720: INFO: (5) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 17.262528ms) -Dec 14 16:22:48.721: INFO: (5) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 17.904718ms) -Dec 14 16:22:48.721: INFO: (5) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test (200; 15.261618ms) -Dec 14 16:22:48.750: INFO: (6) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 16.004571ms) -Dec 14 16:22:48.750: INFO: (6) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 15.723896ms) -Dec 14 16:22:48.750: INFO: (6) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 16.119595ms) -Dec 14 16:22:48.751: INFO: (6) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 17.052464ms) -Dec 14 16:22:48.751: INFO: (6) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 17.210519ms) -Dec 14 16:22:48.753: INFO: (6) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 18.714138ms) -Dec 14 16:22:48.754: INFO: (6) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 19.546697ms) -Dec 14 16:22:48.754: INFO: (6) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 19.754346ms) -Dec 14 16:22:48.754: INFO: (6) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 19.94223ms) -Dec 14 16:22:48.754: INFO: (6) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 20.336436ms) -Dec 14 16:22:48.755: INFO: (6) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 20.40774ms) -Dec 14 16:22:48.756: INFO: (6) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 21.349681ms) -Dec 14 16:22:48.771: INFO: (7) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 13.324638ms) -Dec 14 16:22:48.771: INFO: (7) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 14.711693ms) -Dec 14 16:22:48.771: INFO: (7) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test (200; 15.500755ms) -Dec 14 16:22:48.771: INFO: (7) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 13.996277ms) -Dec 14 16:22:48.772: INFO: (7) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 14.738237ms) -Dec 14 16:22:48.772: INFO: (7) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 15.885285ms) -Dec 14 16:22:48.775: INFO: (7) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 16.957688ms) -Dec 14 16:22:48.776: INFO: (7) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 19.951415ms) -Dec 14 16:22:48.778: INFO: (7) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 19.51758ms) -Dec 14 16:22:48.778: INFO: (7) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 20.299692ms) -Dec 14 16:22:48.780: INFO: (7) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 23.490862ms) -Dec 14 16:22:48.780: INFO: (7) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 21.785346ms) -Dec 14 16:22:48.780: INFO: (7) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 23.366666ms) -Dec 14 16:22:48.780: INFO: (7) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 23.332444ms) -Dec 14 16:22:48.785: INFO: (7) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 26.279376ms) -Dec 14 16:22:48.793: INFO: (8) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 8.518823ms) -Dec 14 16:22:48.795: INFO: (8) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 9.8601ms) -Dec 14 16:22:48.800: INFO: (8) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 14.435683ms) -Dec 14 16:22:48.802: INFO: (8) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 16.940374ms) -Dec 14 16:22:48.802: INFO: (8) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 16.641254ms) -Dec 14 16:22:48.804: INFO: (8) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 18.744286ms) -Dec 14 16:22:48.804: INFO: (8) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 18.938085ms) -Dec 14 16:22:48.805: INFO: (8) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: ... (200; 7.394699ms) -Dec 14 16:22:48.820: INFO: (9) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 7.395498ms) -Dec 14 16:22:48.825: INFO: (9) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 12.380282ms) -Dec 14 16:22:48.826: INFO: (9) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 13.00977ms) -Dec 14 16:22:48.826: INFO: (9) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 12.3687ms) -Dec 14 16:22:48.826: INFO: (9) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 12.513659ms) -Dec 14 16:22:48.831: INFO: (9) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 17.769723ms) -Dec 14 16:22:48.831: INFO: (9) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 17.780338ms) -Dec 14 16:22:48.831: INFO: (9) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 17.98723ms) -Dec 14 16:22:48.831: INFO: (9) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 18.272718ms) -Dec 14 16:22:48.831: INFO: (9) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 18.731134ms) -Dec 14 16:22:48.832: INFO: (9) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 11.52324ms) -Dec 14 16:22:48.855: INFO: (10) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 16.736733ms) -Dec 14 16:22:48.855: INFO: (10) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 17.26268ms) -Dec 14 16:22:48.855: INFO: (10) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 17.397581ms) -Dec 14 16:22:48.856: INFO: (10) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 18.721862ms) -Dec 14 16:22:48.856: INFO: (10) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 18.283628ms) -Dec 14 16:22:48.856: INFO: (10) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 18.995365ms) -Dec 14 16:22:48.856: INFO: (10) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 18.979483ms) -Dec 14 16:22:48.857: INFO: (10) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 19.669923ms) -Dec 14 16:22:48.857: INFO: (10) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 18.765552ms) -Dec 14 16:22:48.857: INFO: (10) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 19.278234ms) -Dec 14 16:22:48.858: INFO: (10) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 19.898681ms) -Dec 14 16:22:48.858: INFO: (10) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 20.374098ms) -Dec 14 16:22:48.870: INFO: (11) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 11.209985ms) -Dec 14 16:22:48.872: INFO: (11) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 13.278985ms) -Dec 14 16:22:48.872: INFO: (11) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 12.415804ms) -Dec 14 16:22:48.872: INFO: (11) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 13.781224ms) -Dec 14 16:22:48.873: INFO: (11) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 14.915756ms) -Dec 14 16:22:48.874: INFO: (11) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 14.478379ms) -Dec 14 16:22:48.875: INFO: (11) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 16.089547ms) -Dec 14 16:22:48.875: INFO: (11) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 16.803898ms) -Dec 14 16:22:48.875: INFO: (11) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 16.856223ms) -Dec 14 16:22:48.875: INFO: (11) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 15.42949ms) -Dec 14 16:22:48.875: INFO: (11) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 16.385207ms) -Dec 14 16:22:48.876: INFO: (11) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 16.211471ms) -Dec 14 16:22:48.880: INFO: (11) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 20.430432ms) -Dec 14 16:22:48.880: INFO: (11) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 19.869742ms) -Dec 14 16:22:48.880: INFO: (11) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 20.776464ms) -Dec 14 16:22:48.885: INFO: (12) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 5.069061ms) -Dec 14 16:22:48.897: INFO: (12) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 17.117711ms) -Dec 14 16:22:48.899: INFO: (12) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 18.5488ms) -Dec 14 16:22:48.899: INFO: (12) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 18.444326ms) -Dec 14 16:22:48.899: INFO: (12) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 18.59111ms) -Dec 14 16:22:48.900: INFO: (12) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 19.971492ms) -Dec 14 16:22:48.900: INFO: (12) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 19.707801ms) -Dec 14 16:22:48.901: INFO: (12) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 20.89554ms) -Dec 14 16:22:48.902: INFO: (12) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 21.04724ms) -Dec 14 16:22:48.902: INFO: (12) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: ... (200; 21.834131ms) -Dec 14 16:22:48.904: INFO: (12) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 23.037872ms) -Dec 14 16:22:48.904: INFO: (12) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 23.100946ms) -Dec 14 16:22:48.904: INFO: (12) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 22.811219ms) -Dec 14 16:22:48.906: INFO: (12) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 25.656501ms) -Dec 14 16:22:48.906: INFO: (12) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 25.699877ms) -Dec 14 16:22:48.914: INFO: (13) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 7.743908ms) -Dec 14 16:22:48.915: INFO: (13) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 7.800246ms) -Dec 14 16:22:48.916: INFO: (13) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 8.653049ms) -Dec 14 16:22:48.917: INFO: (13) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 10.297467ms) -Dec 14 16:22:48.918: INFO: (13) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 10.955083ms) -Dec 14 16:22:48.921: INFO: (13) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 13.176399ms) -Dec 14 16:22:48.921: INFO: (13) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 13.668616ms) -Dec 14 16:22:48.921: INFO: (13) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 14.302593ms) -Dec 14 16:22:48.922: INFO: (13) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 14.009679ms) -Dec 14 16:22:48.923: INFO: (13) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 15.56929ms) -Dec 14 16:22:48.923: INFO: (13) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 13.230869ms) -Dec 14 16:22:48.946: INFO: (14) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 16.128846ms) -Dec 14 16:22:48.947: INFO: (14) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 16.131688ms) -Dec 14 16:22:48.948: INFO: (14) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: ... (200; 19.115841ms) -Dec 14 16:22:48.951: INFO: (14) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 21.463374ms) -Dec 14 16:22:48.952: INFO: (14) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 20.34446ms) -Dec 14 16:22:48.952: INFO: (14) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 21.824086ms) -Dec 14 16:22:48.952: INFO: (14) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 20.243493ms) -Dec 14 16:22:48.953: INFO: (14) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 21.963818ms) -Dec 14 16:22:48.954: INFO: (14) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 23.255493ms) -Dec 14 16:22:48.966: INFO: (15) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 10.690249ms) -Dec 14 16:22:48.966: INFO: (15) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 11.073381ms) -Dec 14 16:22:48.974: INFO: (15) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 18.711499ms) -Dec 14 16:22:48.974: INFO: (15) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 18.994973ms) -Dec 14 16:22:48.977: INFO: (15) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 21.16892ms) -Dec 14 16:22:48.978: INFO: (15) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 22.593222ms) -Dec 14 16:22:48.978: INFO: (15) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 21.55605ms) -Dec 14 16:22:48.979: INFO: (15) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 24.181791ms) -Dec 14 16:22:48.980: INFO: (15) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: ... (200; 24.620925ms) -Dec 14 16:22:48.981: INFO: (15) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 24.362937ms) -Dec 14 16:22:48.981: INFO: (15) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 25.766027ms) -Dec 14 16:22:48.981: INFO: (15) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 26.194226ms) -Dec 14 16:22:48.981: INFO: (15) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 25.712861ms) -Dec 14 16:22:48.982: INFO: (15) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 25.835455ms) -Dec 14 16:22:48.983: INFO: (15) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 26.898206ms) -Dec 14 16:22:48.989: INFO: (16) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 5.881837ms) -Dec 14 16:22:48.989: INFO: (16) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 14.67817ms) -Dec 14 16:22:49.000: INFO: (16) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 16.191957ms) -Dec 14 16:22:49.001: INFO: (16) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 17.857722ms) -Dec 14 16:22:49.002: INFO: (16) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 18.192039ms) -Dec 14 16:22:49.002: INFO: (16) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 18.555605ms) -Dec 14 16:22:49.002: INFO: (16) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 18.342648ms) -Dec 14 16:22:49.002: INFO: (16) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 18.567406ms) -Dec 14 16:22:49.002: INFO: (16) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 18.659583ms) -Dec 14 16:22:49.003: INFO: (16) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 19.668942ms) -Dec 14 16:22:49.003: INFO: (16) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 20.356874ms) -Dec 14 16:22:49.004: INFO: (16) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 20.998499ms) -Dec 14 16:22:49.006: INFO: (16) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 22.901651ms) -Dec 14 16:22:49.007: INFO: (16) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 23.387252ms) -Dec 14 16:22:49.007: INFO: (16) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 23.69476ms) -Dec 14 16:22:49.024: INFO: (17) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 16.459582ms) -Dec 14 16:22:49.025: INFO: (17) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 18.213352ms) -Dec 14 16:22:49.025: INFO: (17) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: ... (200; 23.522615ms) -Dec 14 16:22:49.032: INFO: (17) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 23.872282ms) -Dec 14 16:22:49.032: INFO: (17) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 24.689644ms) -Dec 14 16:22:49.032: INFO: (17) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 24.185944ms) -Dec 14 16:22:49.033: INFO: (17) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 24.354709ms) -Dec 14 16:22:49.035: INFO: (17) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 25.942982ms) -Dec 14 16:22:49.035: INFO: (17) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 27.25971ms) -Dec 14 16:22:49.036: INFO: (17) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 27.59634ms) -Dec 14 16:22:49.038: INFO: (17) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 29.44083ms) -Dec 14 16:22:49.045: INFO: (18) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 6.515994ms) -Dec 14 16:22:49.045: INFO: (18) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 6.991231ms) -Dec 14 16:22:49.048: INFO: (18) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 9.727128ms) -Dec 14 16:22:49.051: INFO: (18) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 12.054871ms) -Dec 14 16:22:49.054: INFO: (18) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 14.993781ms) -Dec 14 16:22:49.054: INFO: (18) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 15.267405ms) -Dec 14 16:22:49.056: INFO: (18) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 16.906866ms) -Dec 14 16:22:49.056: INFO: (18) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 16.799819ms) -Dec 14 16:22:49.056: INFO: (18) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 17.532157ms) -Dec 14 16:22:49.057: INFO: (18) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 17.852554ms) -Dec 14 16:22:49.059: INFO: (18) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 20.06216ms) -Dec 14 16:22:49.059: INFO: (18) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 20.782474ms) -Dec 14 16:22:49.068: INFO: (19) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 8.370362ms) -Dec 14 16:22:49.069: INFO: (19) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 9.724942ms) -Dec 14 16:22:49.069: INFO: (19) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 9.12785ms) -Dec 14 16:22:49.070: INFO: (19) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 43.868921ms) -Dec 14 16:22:49.104: INFO: (19) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 44.346041ms) -Dec 14 16:22:49.105: INFO: (19) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 44.542367ms) -Dec 14 16:22:49.106: INFO: (19) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 45.382739ms) -Dec 14 16:22:49.106: INFO: (19) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 46.198615ms) -STEP: deleting ReplicationController proxy-service-xn49c in namespace proxy-6092, will wait for the garbage collector to delete the pods 12/14/22 16:22:49.106 -Dec 14 16:22:49.178: INFO: Deleting ReplicationController proxy-service-xn49c took: 12.975031ms -Dec 14 16:22:49.279: INFO: Terminating ReplicationController proxy-service-xn49c pods took: 100.367381ms -[AfterEach] version v1 +STEP: Creating a kubernetes client 02/25/23 10:21:36.212 +Feb 25 10:21:36.212: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename dns 02/25/23 10:21:36.22 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:21:36.258 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:21:36.263 +[It] should provide DNS for pods for Hostname [Conformance] + test/e2e/network/dns.go:248 +STEP: Creating a test headless service 02/25/23 10:21:36.27 +STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-9141.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-2.dns-test-service-2.dns-9141.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/wheezy_hosts@dns-querier-2;sleep 1; done + 02/25/23 10:21:36.279 +STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-9141.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-2.dns-test-service-2.dns-9141.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/jessie_hosts@dns-querier-2;sleep 1; done + 02/25/23 10:21:36.279 +STEP: creating a pod to probe DNS 02/25/23 10:21:36.28 +STEP: submitting the pod to kubernetes 02/25/23 10:21:36.28 +Feb 25 10:21:36.306: INFO: Waiting up to 15m0s for pod "dns-test-ca64578f-bd56-4211-a8d6-3b8bccfc4373" in namespace "dns-9141" to be "running" +Feb 25 10:21:36.316: INFO: Pod "dns-test-ca64578f-bd56-4211-a8d6-3b8bccfc4373": Phase="Pending", Reason="", readiness=false. Elapsed: 10.445027ms +Feb 25 10:21:38.330: INFO: Pod "dns-test-ca64578f-bd56-4211-a8d6-3b8bccfc4373": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024676606s +Feb 25 10:21:40.333: INFO: Pod "dns-test-ca64578f-bd56-4211-a8d6-3b8bccfc4373": Phase="Running", Reason="", readiness=true. Elapsed: 4.027329363s +Feb 25 10:21:40.333: INFO: Pod "dns-test-ca64578f-bd56-4211-a8d6-3b8bccfc4373" satisfied condition "running" +STEP: retrieving the pod 02/25/23 10:21:40.333 +STEP: looking for the results for each expected name from probers 02/25/23 10:21:40.349 +Feb 25 10:21:40.405: INFO: DNS probes using dns-9141/dns-test-ca64578f-bd56-4211-a8d6-3b8bccfc4373 succeeded + +STEP: deleting the pod 02/25/23 10:21:40.406 +STEP: deleting the test headless service 02/25/23 10:21:40.456 +[AfterEach] [sig-network] DNS test/e2e/framework/framework.go:187 -Dec 14 16:22:50.680: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "proxy-6092" for this suite. 12/14/22 16:22:50.686 -{"msg":"PASSED [sig-network] Proxy version v1 should proxy through a service and a pod [Conformance]","completed":131,"skipped":2348,"failed":0} +Feb 25 10:21:40.492: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "dns-9141" for this suite. 02/25/23 10:21:40.505 +{"msg":"PASSED [sig-network] DNS should provide DNS for pods for Hostname [Conformance]","completed":125,"skipped":2315,"failed":0} ------------------------------ -• [SLOW TEST] [5.262 seconds] -[sig-network] Proxy +• [4.321 seconds] +[sig-network] DNS test/e2e/network/common/framework.go:23 - version v1 - test/e2e/network/proxy.go:74 - should proxy through a service and a pod [Conformance] - test/e2e/network/proxy.go:101 + should provide DNS for pods for Hostname [Conformance] + test/e2e/network/dns.go:248 Begin Captured GinkgoWriter Output >> - [BeforeEach] version v1 + [BeforeEach] [sig-network] DNS test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:22:45.434 - Dec 14 16:22:45.435: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename proxy 12/14/22 16:22:45.436 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:45.464 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:45.467 - [It] should proxy through a service and a pod [Conformance] - test/e2e/network/proxy.go:101 - STEP: starting an echo server on multiple ports 12/14/22 16:22:45.487 - STEP: creating replication controller proxy-service-xn49c in namespace proxy-6092 12/14/22 16:22:45.487 - I1214 16:22:45.499732 14 runners.go:193] Created replication controller with name: proxy-service-xn49c, namespace: proxy-6092, replica count: 1 - I1214 16:22:46.550851 14 runners.go:193] proxy-service-xn49c Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - I1214 16:22:47.551380 14 runners.go:193] proxy-service-xn49c Pods: 1 out of 1 created, 0 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 1 runningButNotReady - I1214 16:22:48.552422 14 runners.go:193] proxy-service-xn49c Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - Dec 14 16:22:48.560: INFO: setup took 3.088408066s, starting test cases - STEP: running 16 cases, 20 attempts per case, 320 total attempts 12/14/22 16:22:48.56 - Dec 14 16:22:48.574: INFO: (0) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 13.621855ms) - Dec 14 16:22:48.583: INFO: (0) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 23.122439ms) - Dec 14 16:22:48.584: INFO: (0) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 22.864897ms) - Dec 14 16:22:48.584: INFO: (0) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 23.402371ms) - Dec 14 16:22:48.584: INFO: (0) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 22.538269ms) - Dec 14 16:22:48.584: INFO: (0) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 22.516574ms) - Dec 14 16:22:48.584: INFO: (0) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 22.503565ms) - Dec 14 16:22:48.584: INFO: (0) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 22.766861ms) - Dec 14 16:22:48.589: INFO: (0) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 28.430669ms) - Dec 14 16:22:48.589: INFO: (0) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 27.387612ms) - Dec 14 16:22:48.593: INFO: (0) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 31.935466ms) - Dec 14 16:22:48.593: INFO: (0) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 32.474889ms) - Dec 14 16:22:48.593: INFO: (0) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test (200; 12.608289ms) - Dec 14 16:22:48.609: INFO: (1) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 12.283916ms) - Dec 14 16:22:48.609: INFO: (1) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 13.385954ms) - Dec 14 16:22:48.610: INFO: (1) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 13.829023ms) - Dec 14 16:22:48.612: INFO: (1) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 15.284702ms) - Dec 14 16:22:48.613: INFO: (1) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 16.63793ms) - Dec 14 16:22:48.615: INFO: (1) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 18.094382ms) - Dec 14 16:22:48.615: INFO: (1) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 18.700518ms) - Dec 14 16:22:48.615: INFO: (1) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: ... (200; 17.981849ms) - Dec 14 16:22:48.644: INFO: (2) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 19.011543ms) - Dec 14 16:22:48.644: INFO: (2) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 21.556156ms) - Dec 14 16:22:48.648: INFO: (2) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 23.277576ms) - Dec 14 16:22:48.648: INFO: (2) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 23.317741ms) - Dec 14 16:22:48.649: INFO: (2) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 23.675129ms) - Dec 14 16:22:48.649: INFO: (2) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 24.270845ms) - Dec 14 16:22:48.649: INFO: (2) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 24.440642ms) - Dec 14 16:22:48.649: INFO: (2) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 24.135461ms) - Dec 14 16:22:48.649: INFO: (2) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 24.535369ms) - Dec 14 16:22:48.649: INFO: (2) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 23.644196ms) - Dec 14 16:22:48.654: INFO: (2) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 29.532621ms) - Dec 14 16:22:48.655: INFO: (2) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 29.528768ms) - Dec 14 16:22:48.655: INFO: (2) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 30.537116ms) - Dec 14 16:22:48.667: INFO: (3) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 10.888789ms) - Dec 14 16:22:48.668: INFO: (3) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test (200; 14.12124ms) - Dec 14 16:22:48.670: INFO: (3) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 14.094639ms) - Dec 14 16:22:48.670: INFO: (3) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 13.799731ms) - Dec 14 16:22:48.670: INFO: (3) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 13.914116ms) - Dec 14 16:22:48.670: INFO: (3) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 13.783736ms) - Dec 14 16:22:48.671: INFO: (3) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 14.43458ms) - Dec 14 16:22:48.673: INFO: (3) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 17.180541ms) - Dec 14 16:22:48.674: INFO: (3) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 17.249456ms) - Dec 14 16:22:48.675: INFO: (3) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 18.941391ms) - Dec 14 16:22:48.675: INFO: (3) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 18.921959ms) - Dec 14 16:22:48.675: INFO: (3) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 19.103763ms) - Dec 14 16:22:48.676: INFO: (3) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 18.979113ms) - Dec 14 16:22:48.676: INFO: (3) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 19.383251ms) - Dec 14 16:22:48.679: INFO: (3) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 22.379524ms) - Dec 14 16:22:48.692: INFO: (4) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 12.535147ms) - Dec 14 16:22:48.693: INFO: (4) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 13.233753ms) - Dec 14 16:22:48.694: INFO: (4) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 14.63415ms) - Dec 14 16:22:48.694: INFO: (4) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 14.361648ms) - Dec 14 16:22:48.695: INFO: (4) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 14.112028ms) - Dec 14 16:22:48.701: INFO: (4) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 22.162349ms) - Dec 14 16:22:48.701: INFO: (4) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 20.989005ms) - Dec 14 16:22:48.701: INFO: (4) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 21.970152ms) - Dec 14 16:22:48.702: INFO: (4) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 23.214586ms) - Dec 14 16:22:48.702: INFO: (4) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 22.425565ms) - Dec 14 16:22:48.702: INFO: (4) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 16.668387ms) - Dec 14 16:22:48.720: INFO: (5) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 16.48941ms) - Dec 14 16:22:48.720: INFO: (5) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 16.960944ms) - Dec 14 16:22:48.720: INFO: (5) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 17.262528ms) - Dec 14 16:22:48.721: INFO: (5) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 17.904718ms) - Dec 14 16:22:48.721: INFO: (5) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test (200; 15.261618ms) - Dec 14 16:22:48.750: INFO: (6) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 16.004571ms) - Dec 14 16:22:48.750: INFO: (6) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 15.723896ms) - Dec 14 16:22:48.750: INFO: (6) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 16.119595ms) - Dec 14 16:22:48.751: INFO: (6) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 17.052464ms) - Dec 14 16:22:48.751: INFO: (6) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 17.210519ms) - Dec 14 16:22:48.753: INFO: (6) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 18.714138ms) - Dec 14 16:22:48.754: INFO: (6) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 19.546697ms) - Dec 14 16:22:48.754: INFO: (6) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 19.754346ms) - Dec 14 16:22:48.754: INFO: (6) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 19.94223ms) - Dec 14 16:22:48.754: INFO: (6) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 20.336436ms) - Dec 14 16:22:48.755: INFO: (6) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 20.40774ms) - Dec 14 16:22:48.756: INFO: (6) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 21.349681ms) - Dec 14 16:22:48.771: INFO: (7) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 13.324638ms) - Dec 14 16:22:48.771: INFO: (7) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 14.711693ms) - Dec 14 16:22:48.771: INFO: (7) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test (200; 15.500755ms) - Dec 14 16:22:48.771: INFO: (7) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 13.996277ms) - Dec 14 16:22:48.772: INFO: (7) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 14.738237ms) - Dec 14 16:22:48.772: INFO: (7) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 15.885285ms) - Dec 14 16:22:48.775: INFO: (7) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 16.957688ms) - Dec 14 16:22:48.776: INFO: (7) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 19.951415ms) - Dec 14 16:22:48.778: INFO: (7) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 19.51758ms) - Dec 14 16:22:48.778: INFO: (7) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 20.299692ms) - Dec 14 16:22:48.780: INFO: (7) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 23.490862ms) - Dec 14 16:22:48.780: INFO: (7) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 21.785346ms) - Dec 14 16:22:48.780: INFO: (7) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 23.366666ms) - Dec 14 16:22:48.780: INFO: (7) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 23.332444ms) - Dec 14 16:22:48.785: INFO: (7) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 26.279376ms) - Dec 14 16:22:48.793: INFO: (8) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 8.518823ms) - Dec 14 16:22:48.795: INFO: (8) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 9.8601ms) - Dec 14 16:22:48.800: INFO: (8) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 14.435683ms) - Dec 14 16:22:48.802: INFO: (8) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 16.940374ms) - Dec 14 16:22:48.802: INFO: (8) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 16.641254ms) - Dec 14 16:22:48.804: INFO: (8) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 18.744286ms) - Dec 14 16:22:48.804: INFO: (8) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 18.938085ms) - Dec 14 16:22:48.805: INFO: (8) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: ... (200; 7.394699ms) - Dec 14 16:22:48.820: INFO: (9) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 7.395498ms) - Dec 14 16:22:48.825: INFO: (9) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 12.380282ms) - Dec 14 16:22:48.826: INFO: (9) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 13.00977ms) - Dec 14 16:22:48.826: INFO: (9) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 12.3687ms) - Dec 14 16:22:48.826: INFO: (9) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 12.513659ms) - Dec 14 16:22:48.831: INFO: (9) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 17.769723ms) - Dec 14 16:22:48.831: INFO: (9) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 17.780338ms) - Dec 14 16:22:48.831: INFO: (9) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 17.98723ms) - Dec 14 16:22:48.831: INFO: (9) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 18.272718ms) - Dec 14 16:22:48.831: INFO: (9) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 18.731134ms) - Dec 14 16:22:48.832: INFO: (9) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 11.52324ms) - Dec 14 16:22:48.855: INFO: (10) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 16.736733ms) - Dec 14 16:22:48.855: INFO: (10) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 17.26268ms) - Dec 14 16:22:48.855: INFO: (10) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 17.397581ms) - Dec 14 16:22:48.856: INFO: (10) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 18.721862ms) - Dec 14 16:22:48.856: INFO: (10) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 18.283628ms) - Dec 14 16:22:48.856: INFO: (10) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 18.995365ms) - Dec 14 16:22:48.856: INFO: (10) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 18.979483ms) - Dec 14 16:22:48.857: INFO: (10) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 19.669923ms) - Dec 14 16:22:48.857: INFO: (10) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 18.765552ms) - Dec 14 16:22:48.857: INFO: (10) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 19.278234ms) - Dec 14 16:22:48.858: INFO: (10) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 19.898681ms) - Dec 14 16:22:48.858: INFO: (10) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 20.374098ms) - Dec 14 16:22:48.870: INFO: (11) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 11.209985ms) - Dec 14 16:22:48.872: INFO: (11) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 13.278985ms) - Dec 14 16:22:48.872: INFO: (11) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 12.415804ms) - Dec 14 16:22:48.872: INFO: (11) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 13.781224ms) - Dec 14 16:22:48.873: INFO: (11) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 14.915756ms) - Dec 14 16:22:48.874: INFO: (11) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 14.478379ms) - Dec 14 16:22:48.875: INFO: (11) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 16.089547ms) - Dec 14 16:22:48.875: INFO: (11) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 16.803898ms) - Dec 14 16:22:48.875: INFO: (11) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 16.856223ms) - Dec 14 16:22:48.875: INFO: (11) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 15.42949ms) - Dec 14 16:22:48.875: INFO: (11) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 16.385207ms) - Dec 14 16:22:48.876: INFO: (11) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 16.211471ms) - Dec 14 16:22:48.880: INFO: (11) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 20.430432ms) - Dec 14 16:22:48.880: INFO: (11) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 19.869742ms) - Dec 14 16:22:48.880: INFO: (11) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 20.776464ms) - Dec 14 16:22:48.885: INFO: (12) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 5.069061ms) - Dec 14 16:22:48.897: INFO: (12) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 17.117711ms) - Dec 14 16:22:48.899: INFO: (12) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 18.5488ms) - Dec 14 16:22:48.899: INFO: (12) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 18.444326ms) - Dec 14 16:22:48.899: INFO: (12) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 18.59111ms) - Dec 14 16:22:48.900: INFO: (12) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 19.971492ms) - Dec 14 16:22:48.900: INFO: (12) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 19.707801ms) - Dec 14 16:22:48.901: INFO: (12) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 20.89554ms) - Dec 14 16:22:48.902: INFO: (12) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 21.04724ms) - Dec 14 16:22:48.902: INFO: (12) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: ... (200; 21.834131ms) - Dec 14 16:22:48.904: INFO: (12) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 23.037872ms) - Dec 14 16:22:48.904: INFO: (12) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 23.100946ms) - Dec 14 16:22:48.904: INFO: (12) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 22.811219ms) - Dec 14 16:22:48.906: INFO: (12) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 25.656501ms) - Dec 14 16:22:48.906: INFO: (12) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 25.699877ms) - Dec 14 16:22:48.914: INFO: (13) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 7.743908ms) - Dec 14 16:22:48.915: INFO: (13) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 7.800246ms) - Dec 14 16:22:48.916: INFO: (13) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 8.653049ms) - Dec 14 16:22:48.917: INFO: (13) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 10.297467ms) - Dec 14 16:22:48.918: INFO: (13) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 10.955083ms) - Dec 14 16:22:48.921: INFO: (13) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 13.176399ms) - Dec 14 16:22:48.921: INFO: (13) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 13.668616ms) - Dec 14 16:22:48.921: INFO: (13) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 14.302593ms) - Dec 14 16:22:48.922: INFO: (13) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 14.009679ms) - Dec 14 16:22:48.923: INFO: (13) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 15.56929ms) - Dec 14 16:22:48.923: INFO: (13) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 13.230869ms) - Dec 14 16:22:48.946: INFO: (14) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 16.128846ms) - Dec 14 16:22:48.947: INFO: (14) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 16.131688ms) - Dec 14 16:22:48.948: INFO: (14) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: ... (200; 19.115841ms) - Dec 14 16:22:48.951: INFO: (14) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 21.463374ms) - Dec 14 16:22:48.952: INFO: (14) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 20.34446ms) - Dec 14 16:22:48.952: INFO: (14) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 21.824086ms) - Dec 14 16:22:48.952: INFO: (14) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 20.243493ms) - Dec 14 16:22:48.953: INFO: (14) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 21.963818ms) - Dec 14 16:22:48.954: INFO: (14) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 23.255493ms) - Dec 14 16:22:48.966: INFO: (15) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 10.690249ms) - Dec 14 16:22:48.966: INFO: (15) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 11.073381ms) - Dec 14 16:22:48.974: INFO: (15) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 18.711499ms) - Dec 14 16:22:48.974: INFO: (15) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 18.994973ms) - Dec 14 16:22:48.977: INFO: (15) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 21.16892ms) - Dec 14 16:22:48.978: INFO: (15) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 22.593222ms) - Dec 14 16:22:48.978: INFO: (15) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 21.55605ms) - Dec 14 16:22:48.979: INFO: (15) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 24.181791ms) - Dec 14 16:22:48.980: INFO: (15) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: ... (200; 24.620925ms) - Dec 14 16:22:48.981: INFO: (15) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 24.362937ms) - Dec 14 16:22:48.981: INFO: (15) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 25.766027ms) - Dec 14 16:22:48.981: INFO: (15) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 26.194226ms) - Dec 14 16:22:48.981: INFO: (15) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 25.712861ms) - Dec 14 16:22:48.982: INFO: (15) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 25.835455ms) - Dec 14 16:22:48.983: INFO: (15) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 26.898206ms) - Dec 14 16:22:48.989: INFO: (16) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 5.881837ms) - Dec 14 16:22:48.989: INFO: (16) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 14.67817ms) - Dec 14 16:22:49.000: INFO: (16) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 16.191957ms) - Dec 14 16:22:49.001: INFO: (16) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 17.857722ms) - Dec 14 16:22:49.002: INFO: (16) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 18.192039ms) - Dec 14 16:22:49.002: INFO: (16) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 18.555605ms) - Dec 14 16:22:49.002: INFO: (16) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 18.342648ms) - Dec 14 16:22:49.002: INFO: (16) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 18.567406ms) - Dec 14 16:22:49.002: INFO: (16) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 18.659583ms) - Dec 14 16:22:49.003: INFO: (16) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 19.668942ms) - Dec 14 16:22:49.003: INFO: (16) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 20.356874ms) - Dec 14 16:22:49.004: INFO: (16) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 20.998499ms) - Dec 14 16:22:49.006: INFO: (16) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 22.901651ms) - Dec 14 16:22:49.007: INFO: (16) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 23.387252ms) - Dec 14 16:22:49.007: INFO: (16) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 23.69476ms) - Dec 14 16:22:49.024: INFO: (17) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:1080/proxy/: test<... (200; 16.459582ms) - Dec 14 16:22:49.025: INFO: (17) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 18.213352ms) - Dec 14 16:22:49.025: INFO: (17) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: ... (200; 23.522615ms) - Dec 14 16:22:49.032: INFO: (17) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 23.872282ms) - Dec 14 16:22:49.032: INFO: (17) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 24.689644ms) - Dec 14 16:22:49.032: INFO: (17) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 24.185944ms) - Dec 14 16:22:49.033: INFO: (17) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 24.354709ms) - Dec 14 16:22:49.035: INFO: (17) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 25.942982ms) - Dec 14 16:22:49.035: INFO: (17) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname2/proxy/: tls qux (200; 27.25971ms) - Dec 14 16:22:49.036: INFO: (17) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 27.59634ms) - Dec 14 16:22:49.038: INFO: (17) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 29.44083ms) - Dec 14 16:22:49.045: INFO: (18) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname2/proxy/: bar (200; 6.515994ms) - Dec 14 16:22:49.045: INFO: (18) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 6.991231ms) - Dec 14 16:22:49.048: INFO: (18) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 9.727128ms) - Dec 14 16:22:49.051: INFO: (18) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 12.054871ms) - Dec 14 16:22:49.054: INFO: (18) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 14.993781ms) - Dec 14 16:22:49.054: INFO: (18) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 15.267405ms) - Dec 14 16:22:49.056: INFO: (18) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:162/proxy/: bar (200; 16.906866ms) - Dec 14 16:22:49.056: INFO: (18) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 16.799819ms) - Dec 14 16:22:49.056: INFO: (18) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 17.532157ms) - Dec 14 16:22:49.057: INFO: (18) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname1/proxy/: foo (200; 17.852554ms) - Dec 14 16:22:49.059: INFO: (18) /api/v1/namespaces/proxy-6092/services/https:proxy-service-xn49c:tlsportname1/proxy/: tls baz (200; 20.06216ms) - Dec 14 16:22:49.059: INFO: (18) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 20.782474ms) - Dec 14 16:22:49.068: INFO: (19) /api/v1/namespaces/proxy-6092/pods/http:proxy-service-xn49c-lbrw8:1080/proxy/: ... (200; 8.370362ms) - Dec 14 16:22:49.069: INFO: (19) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8:160/proxy/: foo (200; 9.724942ms) - Dec 14 16:22:49.069: INFO: (19) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:462/proxy/: tls qux (200; 9.12785ms) - Dec 14 16:22:49.070: INFO: (19) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:443/proxy/: test<... (200; 43.868921ms) - Dec 14 16:22:49.104: INFO: (19) /api/v1/namespaces/proxy-6092/pods/proxy-service-xn49c-lbrw8/proxy/: test (200; 44.346041ms) - Dec 14 16:22:49.105: INFO: (19) /api/v1/namespaces/proxy-6092/pods/https:proxy-service-xn49c-lbrw8:460/proxy/: tls baz (200; 44.542367ms) - Dec 14 16:22:49.106: INFO: (19) /api/v1/namespaces/proxy-6092/services/proxy-service-xn49c:portname2/proxy/: bar (200; 45.382739ms) - Dec 14 16:22:49.106: INFO: (19) /api/v1/namespaces/proxy-6092/services/http:proxy-service-xn49c:portname1/proxy/: foo (200; 46.198615ms) - STEP: deleting ReplicationController proxy-service-xn49c in namespace proxy-6092, will wait for the garbage collector to delete the pods 12/14/22 16:22:49.106 - Dec 14 16:22:49.178: INFO: Deleting ReplicationController proxy-service-xn49c took: 12.975031ms - Dec 14 16:22:49.279: INFO: Terminating ReplicationController proxy-service-xn49c pods took: 100.367381ms - [AfterEach] version v1 + STEP: Creating a kubernetes client 02/25/23 10:21:36.212 + Feb 25 10:21:36.212: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename dns 02/25/23 10:21:36.22 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:21:36.258 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:21:36.263 + [It] should provide DNS for pods for Hostname [Conformance] + test/e2e/network/dns.go:248 + STEP: Creating a test headless service 02/25/23 10:21:36.27 + STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-9141.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-2.dns-test-service-2.dns-9141.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/wheezy_hosts@dns-querier-2;sleep 1; done + 02/25/23 10:21:36.279 + STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-2.dns-test-service-2.dns-9141.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-2.dns-test-service-2.dns-9141.svc.cluster.local;test -n "$$(getent hosts dns-querier-2)" && echo OK > /results/jessie_hosts@dns-querier-2;sleep 1; done + 02/25/23 10:21:36.279 + STEP: creating a pod to probe DNS 02/25/23 10:21:36.28 + STEP: submitting the pod to kubernetes 02/25/23 10:21:36.28 + Feb 25 10:21:36.306: INFO: Waiting up to 15m0s for pod "dns-test-ca64578f-bd56-4211-a8d6-3b8bccfc4373" in namespace "dns-9141" to be "running" + Feb 25 10:21:36.316: INFO: Pod "dns-test-ca64578f-bd56-4211-a8d6-3b8bccfc4373": Phase="Pending", Reason="", readiness=false. Elapsed: 10.445027ms + Feb 25 10:21:38.330: INFO: Pod "dns-test-ca64578f-bd56-4211-a8d6-3b8bccfc4373": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024676606s + Feb 25 10:21:40.333: INFO: Pod "dns-test-ca64578f-bd56-4211-a8d6-3b8bccfc4373": Phase="Running", Reason="", readiness=true. Elapsed: 4.027329363s + Feb 25 10:21:40.333: INFO: Pod "dns-test-ca64578f-bd56-4211-a8d6-3b8bccfc4373" satisfied condition "running" + STEP: retrieving the pod 02/25/23 10:21:40.333 + STEP: looking for the results for each expected name from probers 02/25/23 10:21:40.349 + Feb 25 10:21:40.405: INFO: DNS probes using dns-9141/dns-test-ca64578f-bd56-4211-a8d6-3b8bccfc4373 succeeded + + STEP: deleting the pod 02/25/23 10:21:40.406 + STEP: deleting the test headless service 02/25/23 10:21:40.456 + [AfterEach] [sig-network] DNS test/e2e/framework/framework.go:187 - Dec 14 16:22:50.680: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "proxy-6092" for this suite. 12/14/22 16:22:50.686 + Feb 25 10:21:40.492: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "dns-9141" for this suite. 02/25/23 10:21:40.505 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:96 -[BeforeEach] [sig-storage] EmptyDir volumes +[sig-scheduling] SchedulerPredicates [Serial] + validates resource limits of pods that are allowed to run [Conformance] + test/e2e/scheduling/predicates.go:326 +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:22:50.725 -Dec 14 16:22:50.725: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir 12/14/22 16:22:50.727 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:50.752 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:50.756 -[It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:96 -STEP: Creating a pod to test emptydir 0644 on tmpfs 12/14/22 16:22:50.759 -Dec 14 16:22:50.772: INFO: Waiting up to 5m0s for pod "pod-17db4006-e97e-4bdc-8909-b37753370c1f" in namespace "emptydir-8096" to be "Succeeded or Failed" -Dec 14 16:22:50.776: INFO: Pod "pod-17db4006-e97e-4bdc-8909-b37753370c1f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.115149ms -Dec 14 16:22:52.785: INFO: Pod "pod-17db4006-e97e-4bdc-8909-b37753370c1f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013150087s -Dec 14 16:22:54.785: INFO: Pod "pod-17db4006-e97e-4bdc-8909-b37753370c1f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013084603s -STEP: Saw pod success 12/14/22 16:22:54.785 -Dec 14 16:22:54.785: INFO: Pod "pod-17db4006-e97e-4bdc-8909-b37753370c1f" satisfied condition "Succeeded or Failed" -Dec 14 16:22:54.789: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-17db4006-e97e-4bdc-8909-b37753370c1f container test-container: -STEP: delete the pod 12/14/22 16:22:54.801 -Dec 14 16:22:54.821: INFO: Waiting for pod pod-17db4006-e97e-4bdc-8909-b37753370c1f to disappear -Dec 14 16:22:54.824: INFO: Pod pod-17db4006-e97e-4bdc-8909-b37753370c1f no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +STEP: Creating a kubernetes client 02/25/23 10:21:40.559 +Feb 25 10:21:40.559: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename sched-pred 02/25/23 10:21:40.564 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:21:40.607 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:21:40.613 +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:92 +Feb 25 10:21:40.624: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready +Feb 25 10:21:40.665: INFO: Waiting for terminating namespaces to be deleted... +Feb 25 10:21:40.679: INFO: +Logging pods the apiserver thinks is on node diek3dua9gay-1 before test +Feb 25 10:21:40.700: INFO: echo-other-node-cd69fcf6b-hsnk9 from cilium-test started at 2023-02-25 09:38:06 +0000 UTC (2 container statuses recorded) +Feb 25 10:21:40.700: INFO: Container dns-test-server ready: true, restart count 0 +Feb 25 10:21:40.700: INFO: Container echo-other-node ready: true, restart count 0 +Feb 25 10:21:40.700: INFO: cilium-node-init-s84z5 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.700: INFO: Container node-init ready: true, restart count 0 +Feb 25 10:21:40.701: INFO: cilium-r4lc4 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.701: INFO: Container cilium-agent ready: true, restart count 0 +Feb 25 10:21:40.701: INFO: coredns-565d847f94-qqvf6 from kube-system started at 2023-02-25 08:49:06 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.701: INFO: Container coredns ready: true, restart count 0 +Feb 25 10:21:40.701: INFO: kube-addon-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.701: INFO: Container kube-addon-manager ready: true, restart count 0 +Feb 25 10:21:40.701: INFO: kube-apiserver-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:09 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.701: INFO: Container kube-apiserver ready: true, restart count 0 +Feb 25 10:21:40.701: INFO: kube-controller-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:49:55 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.701: INFO: Container kube-controller-manager ready: true, restart count 0 +Feb 25 10:21:40.701: INFO: kube-proxy-cm4qg from kube-system started at 2023-02-25 08:46:38 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.702: INFO: Container kube-proxy ready: true, restart count 0 +Feb 25 10:21:40.702: INFO: kube-scheduler-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:27 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.702: INFO: Container kube-scheduler ready: true, restart count 0 +Feb 25 10:21:40.702: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-wznrx from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 10:21:40.702: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 10:21:40.702: INFO: Container systemd-logs ready: true, restart count 0 +Feb 25 10:21:40.702: INFO: +Logging pods the apiserver thinks is on node diek3dua9gay-2 before test +Feb 25 10:21:40.757: INFO: cilium-9tkfg from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.757: INFO: Container cilium-agent ready: true, restart count 0 +Feb 25 10:21:40.757: INFO: cilium-node-init-8b7n2 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.757: INFO: Container node-init ready: true, restart count 0 +Feb 25 10:21:40.757: INFO: cilium-operator-59d7d77cdc-lppbd from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.757: INFO: Container cilium-operator ready: true, restart count 0 +Feb 25 10:21:40.757: INFO: coredns-565d847f94-5nh2n from kube-system started at 2023-02-25 08:48:51 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.757: INFO: Container coredns ready: true, restart count 0 +Feb 25 10:21:40.757: INFO: kube-addon-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.757: INFO: Container kube-addon-manager ready: true, restart count 0 +Feb 25 10:21:40.757: INFO: kube-apiserver-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.758: INFO: Container kube-apiserver ready: true, restart count 0 +Feb 25 10:21:40.758: INFO: kube-controller-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.758: INFO: Container kube-controller-manager ready: true, restart count 0 +Feb 25 10:21:40.758: INFO: kube-proxy-d66m7 from kube-system started at 2023-02-25 08:46:57 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.758: INFO: Container kube-proxy ready: true, restart count 0 +Feb 25 10:21:40.758: INFO: kube-scheduler-diek3dua9gay-2 from kube-system started at 2023-02-25 08:50:15 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.758: INFO: Container kube-scheduler ready: true, restart count 0 +Feb 25 10:21:40.758: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-mml8w from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 10:21:40.758: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 10:21:40.758: INFO: Container systemd-logs ready: true, restart count 0 +Feb 25 10:21:40.758: INFO: +Logging pods the apiserver thinks is on node diek3dua9gay-3 before test +Feb 25 10:21:40.780: INFO: client-7b78db77d5-27xbt from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.780: INFO: Container client ready: true, restart count 0 +Feb 25 10:21:40.780: INFO: client2-78f748dd67-skwz2 from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.780: INFO: Container client2 ready: true, restart count 0 +Feb 25 10:21:40.780: INFO: echo-same-node-85bc9b6b56-wqmfq from cilium-test started at 2023-02-25 09:57:46 +0000 UTC (2 container statuses recorded) +Feb 25 10:21:40.780: INFO: Container dns-test-server ready: true, restart count 0 +Feb 25 10:21:40.780: INFO: Container echo-same-node ready: true, restart count 0 +Feb 25 10:21:40.780: INFO: cilium-2cncz from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.780: INFO: Container cilium-agent ready: true, restart count 0 +Feb 25 10:21:40.780: INFO: cilium-node-init-979mb from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.780: INFO: Container node-init ready: true, restart count 0 +Feb 25 10:21:40.780: INFO: kube-proxy-22s9q from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.780: INFO: Container kube-proxy ready: true, restart count 0 +Feb 25 10:21:40.780: INFO: sonobuoy from sonobuoy started at 2023-02-25 09:45:14 +0000 UTC (1 container statuses recorded) +Feb 25 10:21:40.781: INFO: Container kube-sonobuoy ready: true, restart count 0 +Feb 25 10:21:40.781: INFO: sonobuoy-e2e-job-0b3c30975b5c4880 from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 10:21:40.781: INFO: Container e2e ready: true, restart count 0 +Feb 25 10:21:40.781: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 10:21:40.781: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-9pdjb from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 10:21:40.781: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 10:21:40.781: INFO: Container systemd-logs ready: true, restart count 0 +[It] validates resource limits of pods that are allowed to run [Conformance] + test/e2e/scheduling/predicates.go:326 +STEP: verifying the node has the label node diek3dua9gay-1 02/25/23 10:21:40.857 +STEP: verifying the node has the label node diek3dua9gay-2 02/25/23 10:21:40.963 +STEP: verifying the node has the label node diek3dua9gay-3 02/25/23 10:21:41.052 +Feb 25 10:21:41.179: INFO: Pod client-7b78db77d5-27xbt requesting resource cpu=0m on Node diek3dua9gay-3 +Feb 25 10:21:41.179: INFO: Pod client2-78f748dd67-skwz2 requesting resource cpu=0m on Node diek3dua9gay-3 +Feb 25 10:21:41.179: INFO: Pod echo-other-node-cd69fcf6b-hsnk9 requesting resource cpu=0m on Node diek3dua9gay-1 +Feb 25 10:21:41.179: INFO: Pod echo-same-node-85bc9b6b56-wqmfq requesting resource cpu=0m on Node diek3dua9gay-3 +Feb 25 10:21:41.179: INFO: Pod cilium-2cncz requesting resource cpu=0m on Node diek3dua9gay-3 +Feb 25 10:21:41.179: INFO: Pod cilium-9tkfg requesting resource cpu=0m on Node diek3dua9gay-2 +Feb 25 10:21:41.179: INFO: Pod cilium-node-init-8b7n2 requesting resource cpu=100m on Node diek3dua9gay-2 +Feb 25 10:21:41.179: INFO: Pod cilium-node-init-979mb requesting resource cpu=100m on Node diek3dua9gay-3 +Feb 25 10:21:41.179: INFO: Pod cilium-node-init-s84z5 requesting resource cpu=100m on Node diek3dua9gay-1 +Feb 25 10:21:41.179: INFO: Pod cilium-operator-59d7d77cdc-lppbd requesting resource cpu=0m on Node diek3dua9gay-2 +Feb 25 10:21:41.179: INFO: Pod cilium-r4lc4 requesting resource cpu=0m on Node diek3dua9gay-1 +Feb 25 10:21:41.179: INFO: Pod coredns-565d847f94-5nh2n requesting resource cpu=100m on Node diek3dua9gay-2 +Feb 25 10:21:41.179: INFO: Pod coredns-565d847f94-qqvf6 requesting resource cpu=100m on Node diek3dua9gay-1 +Feb 25 10:21:41.179: INFO: Pod kube-addon-manager-diek3dua9gay-1 requesting resource cpu=5m on Node diek3dua9gay-1 +Feb 25 10:21:41.179: INFO: Pod kube-addon-manager-diek3dua9gay-2 requesting resource cpu=5m on Node diek3dua9gay-2 +Feb 25 10:21:41.179: INFO: Pod kube-apiserver-diek3dua9gay-1 requesting resource cpu=250m on Node diek3dua9gay-1 +Feb 25 10:21:41.179: INFO: Pod kube-apiserver-diek3dua9gay-2 requesting resource cpu=250m on Node diek3dua9gay-2 +Feb 25 10:21:41.179: INFO: Pod kube-controller-manager-diek3dua9gay-1 requesting resource cpu=200m on Node diek3dua9gay-1 +Feb 25 10:21:41.179: INFO: Pod kube-controller-manager-diek3dua9gay-2 requesting resource cpu=200m on Node diek3dua9gay-2 +Feb 25 10:21:41.179: INFO: Pod kube-proxy-22s9q requesting resource cpu=0m on Node diek3dua9gay-3 +Feb 25 10:21:41.179: INFO: Pod kube-proxy-cm4qg requesting resource cpu=0m on Node diek3dua9gay-1 +Feb 25 10:21:41.179: INFO: Pod kube-proxy-d66m7 requesting resource cpu=0m on Node diek3dua9gay-2 +Feb 25 10:21:41.179: INFO: Pod kube-scheduler-diek3dua9gay-1 requesting resource cpu=100m on Node diek3dua9gay-1 +Feb 25 10:21:41.179: INFO: Pod kube-scheduler-diek3dua9gay-2 requesting resource cpu=100m on Node diek3dua9gay-2 +Feb 25 10:21:41.179: INFO: Pod sonobuoy requesting resource cpu=0m on Node diek3dua9gay-3 +Feb 25 10:21:41.179: INFO: Pod sonobuoy-e2e-job-0b3c30975b5c4880 requesting resource cpu=0m on Node diek3dua9gay-3 +Feb 25 10:21:41.179: INFO: Pod sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-9pdjb requesting resource cpu=0m on Node diek3dua9gay-3 +Feb 25 10:21:41.179: INFO: Pod sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-mml8w requesting resource cpu=0m on Node diek3dua9gay-2 +Feb 25 10:21:41.179: INFO: Pod sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-wznrx requesting resource cpu=0m on Node diek3dua9gay-1 +STEP: Starting Pods to consume most of the cluster CPU. 02/25/23 10:21:41.179 +Feb 25 10:21:41.179: INFO: Creating a pod which consumes cpu=591m on Node diek3dua9gay-1 +Feb 25 10:21:41.460: INFO: Creating a pod which consumes cpu=591m on Node diek3dua9gay-2 +Feb 25 10:21:41.631: INFO: Creating a pod which consumes cpu=1050m on Node diek3dua9gay-3 +Feb 25 10:21:41.898: INFO: Waiting up to 5m0s for pod "filler-pod-b8a42890-7232-410e-bc1d-0832f167939b" in namespace "sched-pred-7978" to be "running" +Feb 25 10:21:42.361: INFO: Pod "filler-pod-b8a42890-7232-410e-bc1d-0832f167939b": Phase="Pending", Reason="", readiness=false. Elapsed: 462.663176ms +Feb 25 10:21:44.778: INFO: Pod "filler-pod-b8a42890-7232-410e-bc1d-0832f167939b": Phase="Running", Reason="", readiness=true. Elapsed: 2.879304518s +Feb 25 10:21:44.778: INFO: Pod "filler-pod-b8a42890-7232-410e-bc1d-0832f167939b" satisfied condition "running" +Feb 25 10:21:44.778: INFO: Waiting up to 5m0s for pod "filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a" in namespace "sched-pred-7978" to be "running" +Feb 25 10:21:44.823: INFO: Pod "filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a": Phase="Running", Reason="", readiness=true. Elapsed: 45.187474ms +Feb 25 10:21:45.069: INFO: Pod "filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a" satisfied condition "running" +Feb 25 10:21:45.069: INFO: Waiting up to 5m0s for pod "filler-pod-ecc20828-f391-4389-a241-36eb17240621" in namespace "sched-pred-7978" to be "running" +Feb 25 10:21:45.134: INFO: Pod "filler-pod-ecc20828-f391-4389-a241-36eb17240621": Phase="Pending", Reason="", readiness=false. Elapsed: 64.38347ms +Feb 25 10:21:47.144: INFO: Pod "filler-pod-ecc20828-f391-4389-a241-36eb17240621": Phase="Running", Reason="", readiness=true. Elapsed: 2.074499355s +Feb 25 10:21:47.144: INFO: Pod "filler-pod-ecc20828-f391-4389-a241-36eb17240621" satisfied condition "running" +STEP: Creating another pod that requires unavailable amount of CPU. 02/25/23 10:21:47.144 +STEP: Considering event: +Type = [Normal], Name = [filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a.17470a1c0353a83a], Reason = [Scheduled], Message = [Successfully assigned sched-pred-7978/filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a to diek3dua9gay-2] 02/25/23 10:21:47.153 +STEP: Considering event: +Type = [Normal], Name = [filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a.17470a1c5c70a2d1], Reason = [Pulled], Message = [Container image "registry.k8s.io/pause:3.8" already present on machine] 02/25/23 10:21:47.155 +STEP: Considering event: +Type = [Normal], Name = [filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a.17470a1c77b9d8a0], Reason = [Created], Message = [Created container filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a] 02/25/23 10:21:47.156 +STEP: Considering event: +Type = [Normal], Name = [filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a.17470a1c7a598e1c], Reason = [Started], Message = [Started container filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a] 02/25/23 10:21:47.156 +STEP: Considering event: +Type = [Normal], Name = [filler-pod-b8a42890-7232-410e-bc1d-0832f167939b.17470a1c012bb488], Reason = [Scheduled], Message = [Successfully assigned sched-pred-7978/filler-pod-b8a42890-7232-410e-bc1d-0832f167939b to diek3dua9gay-1] 02/25/23 10:21:47.156 +STEP: Considering event: +Type = [Normal], Name = [filler-pod-b8a42890-7232-410e-bc1d-0832f167939b.17470a1c75a411f8], Reason = [Pulled], Message = [Container image "registry.k8s.io/pause:3.8" already present on machine] 02/25/23 10:21:47.156 +STEP: Considering event: +Type = [Normal], Name = [filler-pod-b8a42890-7232-410e-bc1d-0832f167939b.17470a1c9195d22c], Reason = [Created], Message = [Created container filler-pod-b8a42890-7232-410e-bc1d-0832f167939b] 02/25/23 10:21:47.157 +STEP: Considering event: +Type = [Normal], Name = [filler-pod-b8a42890-7232-410e-bc1d-0832f167939b.17470a1c97cabc5b], Reason = [Started], Message = [Started container filler-pod-b8a42890-7232-410e-bc1d-0832f167939b] 02/25/23 10:21:47.157 +STEP: Considering event: +Type = [Normal], Name = [filler-pod-ecc20828-f391-4389-a241-36eb17240621.17470a1c0dad586a], Reason = [Scheduled], Message = [Successfully assigned sched-pred-7978/filler-pod-ecc20828-f391-4389-a241-36eb17240621 to diek3dua9gay-3] 02/25/23 10:21:47.158 +STEP: Considering event: +Type = [Normal], Name = [filler-pod-ecc20828-f391-4389-a241-36eb17240621.17470a1cdfc1a099], Reason = [Pulled], Message = [Container image "registry.k8s.io/pause:3.8" already present on machine] 02/25/23 10:21:47.158 +STEP: Considering event: +Type = [Normal], Name = [filler-pod-ecc20828-f391-4389-a241-36eb17240621.17470a1d0fe43b6b], Reason = [Created], Message = [Created container filler-pod-ecc20828-f391-4389-a241-36eb17240621] 02/25/23 10:21:47.158 +STEP: Considering event: +Type = [Normal], Name = [filler-pod-ecc20828-f391-4389-a241-36eb17240621.17470a1d1489bff1], Reason = [Started], Message = [Started container filler-pod-ecc20828-f391-4389-a241-36eb17240621] 02/25/23 10:21:47.158 +STEP: Considering event: +Type = [Warning], Name = [additional-pod.17470a1d4df2f4de], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 Insufficient cpu. preemption: 0/3 nodes are available: 3 No preemption victims found for incoming pod.] 02/25/23 10:21:47.202 +STEP: removing the label node off the node diek3dua9gay-1 02/25/23 10:21:48.181 +STEP: verifying the node doesn't have the label node 02/25/23 10:21:48.219 +STEP: removing the label node off the node diek3dua9gay-2 02/25/23 10:21:48.228 +STEP: verifying the node doesn't have the label node 02/25/23 10:21:48.252 +STEP: removing the label node off the node diek3dua9gay-3 02/25/23 10:21:48.262 +STEP: verifying the node doesn't have the label node 02/25/23 10:21:48.293 +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:187 -Dec 14 16:22:54.825: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-8096" for this suite. 12/14/22 16:22:54.831 -{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","completed":132,"skipped":2378,"failed":0} +Feb 25 10:21:48.317: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "sched-pred-7978" for this suite. 02/25/23 10:21:48.333 +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:83 +{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates resource limits of pods that are allowed to run [Conformance]","completed":126,"skipped":2344,"failed":0} ------------------------------ -• [4.114 seconds] -[sig-storage] EmptyDir volumes -test/e2e/common/storage/framework.go:23 - should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:96 +• [SLOW TEST] [7.793 seconds] +[sig-scheduling] SchedulerPredicates [Serial] +test/e2e/scheduling/framework.go:40 + validates resource limits of pods that are allowed to run [Conformance] + test/e2e/scheduling/predicates.go:326 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir volumes + [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:22:50.725 - Dec 14 16:22:50.725: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir 12/14/22 16:22:50.727 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:50.752 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:50.756 - [It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:96 - STEP: Creating a pod to test emptydir 0644 on tmpfs 12/14/22 16:22:50.759 - Dec 14 16:22:50.772: INFO: Waiting up to 5m0s for pod "pod-17db4006-e97e-4bdc-8909-b37753370c1f" in namespace "emptydir-8096" to be "Succeeded or Failed" - Dec 14 16:22:50.776: INFO: Pod "pod-17db4006-e97e-4bdc-8909-b37753370c1f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.115149ms - Dec 14 16:22:52.785: INFO: Pod "pod-17db4006-e97e-4bdc-8909-b37753370c1f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013150087s - Dec 14 16:22:54.785: INFO: Pod "pod-17db4006-e97e-4bdc-8909-b37753370c1f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013084603s - STEP: Saw pod success 12/14/22 16:22:54.785 - Dec 14 16:22:54.785: INFO: Pod "pod-17db4006-e97e-4bdc-8909-b37753370c1f" satisfied condition "Succeeded or Failed" - Dec 14 16:22:54.789: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-17db4006-e97e-4bdc-8909-b37753370c1f container test-container: - STEP: delete the pod 12/14/22 16:22:54.801 - Dec 14 16:22:54.821: INFO: Waiting for pod pod-17db4006-e97e-4bdc-8909-b37753370c1f to disappear - Dec 14 16:22:54.824: INFO: Pod pod-17db4006-e97e-4bdc-8909-b37753370c1f no longer exists - [AfterEach] [sig-storage] EmptyDir volumes + STEP: Creating a kubernetes client 02/25/23 10:21:40.559 + Feb 25 10:21:40.559: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename sched-pred 02/25/23 10:21:40.564 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:21:40.607 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:21:40.613 + [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:92 + Feb 25 10:21:40.624: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready + Feb 25 10:21:40.665: INFO: Waiting for terminating namespaces to be deleted... + Feb 25 10:21:40.679: INFO: + Logging pods the apiserver thinks is on node diek3dua9gay-1 before test + Feb 25 10:21:40.700: INFO: echo-other-node-cd69fcf6b-hsnk9 from cilium-test started at 2023-02-25 09:38:06 +0000 UTC (2 container statuses recorded) + Feb 25 10:21:40.700: INFO: Container dns-test-server ready: true, restart count 0 + Feb 25 10:21:40.700: INFO: Container echo-other-node ready: true, restart count 0 + Feb 25 10:21:40.700: INFO: cilium-node-init-s84z5 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.700: INFO: Container node-init ready: true, restart count 0 + Feb 25 10:21:40.701: INFO: cilium-r4lc4 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.701: INFO: Container cilium-agent ready: true, restart count 0 + Feb 25 10:21:40.701: INFO: coredns-565d847f94-qqvf6 from kube-system started at 2023-02-25 08:49:06 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.701: INFO: Container coredns ready: true, restart count 0 + Feb 25 10:21:40.701: INFO: kube-addon-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.701: INFO: Container kube-addon-manager ready: true, restart count 0 + Feb 25 10:21:40.701: INFO: kube-apiserver-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:09 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.701: INFO: Container kube-apiserver ready: true, restart count 0 + Feb 25 10:21:40.701: INFO: kube-controller-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:49:55 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.701: INFO: Container kube-controller-manager ready: true, restart count 0 + Feb 25 10:21:40.701: INFO: kube-proxy-cm4qg from kube-system started at 2023-02-25 08:46:38 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.702: INFO: Container kube-proxy ready: true, restart count 0 + Feb 25 10:21:40.702: INFO: kube-scheduler-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:27 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.702: INFO: Container kube-scheduler ready: true, restart count 0 + Feb 25 10:21:40.702: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-wznrx from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 10:21:40.702: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 10:21:40.702: INFO: Container systemd-logs ready: true, restart count 0 + Feb 25 10:21:40.702: INFO: + Logging pods the apiserver thinks is on node diek3dua9gay-2 before test + Feb 25 10:21:40.757: INFO: cilium-9tkfg from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.757: INFO: Container cilium-agent ready: true, restart count 0 + Feb 25 10:21:40.757: INFO: cilium-node-init-8b7n2 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.757: INFO: Container node-init ready: true, restart count 0 + Feb 25 10:21:40.757: INFO: cilium-operator-59d7d77cdc-lppbd from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.757: INFO: Container cilium-operator ready: true, restart count 0 + Feb 25 10:21:40.757: INFO: coredns-565d847f94-5nh2n from kube-system started at 2023-02-25 08:48:51 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.757: INFO: Container coredns ready: true, restart count 0 + Feb 25 10:21:40.757: INFO: kube-addon-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.757: INFO: Container kube-addon-manager ready: true, restart count 0 + Feb 25 10:21:40.757: INFO: kube-apiserver-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.758: INFO: Container kube-apiserver ready: true, restart count 0 + Feb 25 10:21:40.758: INFO: kube-controller-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.758: INFO: Container kube-controller-manager ready: true, restart count 0 + Feb 25 10:21:40.758: INFO: kube-proxy-d66m7 from kube-system started at 2023-02-25 08:46:57 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.758: INFO: Container kube-proxy ready: true, restart count 0 + Feb 25 10:21:40.758: INFO: kube-scheduler-diek3dua9gay-2 from kube-system started at 2023-02-25 08:50:15 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.758: INFO: Container kube-scheduler ready: true, restart count 0 + Feb 25 10:21:40.758: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-mml8w from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 10:21:40.758: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 10:21:40.758: INFO: Container systemd-logs ready: true, restart count 0 + Feb 25 10:21:40.758: INFO: + Logging pods the apiserver thinks is on node diek3dua9gay-3 before test + Feb 25 10:21:40.780: INFO: client-7b78db77d5-27xbt from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.780: INFO: Container client ready: true, restart count 0 + Feb 25 10:21:40.780: INFO: client2-78f748dd67-skwz2 from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.780: INFO: Container client2 ready: true, restart count 0 + Feb 25 10:21:40.780: INFO: echo-same-node-85bc9b6b56-wqmfq from cilium-test started at 2023-02-25 09:57:46 +0000 UTC (2 container statuses recorded) + Feb 25 10:21:40.780: INFO: Container dns-test-server ready: true, restart count 0 + Feb 25 10:21:40.780: INFO: Container echo-same-node ready: true, restart count 0 + Feb 25 10:21:40.780: INFO: cilium-2cncz from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.780: INFO: Container cilium-agent ready: true, restart count 0 + Feb 25 10:21:40.780: INFO: cilium-node-init-979mb from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.780: INFO: Container node-init ready: true, restart count 0 + Feb 25 10:21:40.780: INFO: kube-proxy-22s9q from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.780: INFO: Container kube-proxy ready: true, restart count 0 + Feb 25 10:21:40.780: INFO: sonobuoy from sonobuoy started at 2023-02-25 09:45:14 +0000 UTC (1 container statuses recorded) + Feb 25 10:21:40.781: INFO: Container kube-sonobuoy ready: true, restart count 0 + Feb 25 10:21:40.781: INFO: sonobuoy-e2e-job-0b3c30975b5c4880 from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 10:21:40.781: INFO: Container e2e ready: true, restart count 0 + Feb 25 10:21:40.781: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 10:21:40.781: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-9pdjb from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 10:21:40.781: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 10:21:40.781: INFO: Container systemd-logs ready: true, restart count 0 + [It] validates resource limits of pods that are allowed to run [Conformance] + test/e2e/scheduling/predicates.go:326 + STEP: verifying the node has the label node diek3dua9gay-1 02/25/23 10:21:40.857 + STEP: verifying the node has the label node diek3dua9gay-2 02/25/23 10:21:40.963 + STEP: verifying the node has the label node diek3dua9gay-3 02/25/23 10:21:41.052 + Feb 25 10:21:41.179: INFO: Pod client-7b78db77d5-27xbt requesting resource cpu=0m on Node diek3dua9gay-3 + Feb 25 10:21:41.179: INFO: Pod client2-78f748dd67-skwz2 requesting resource cpu=0m on Node diek3dua9gay-3 + Feb 25 10:21:41.179: INFO: Pod echo-other-node-cd69fcf6b-hsnk9 requesting resource cpu=0m on Node diek3dua9gay-1 + Feb 25 10:21:41.179: INFO: Pod echo-same-node-85bc9b6b56-wqmfq requesting resource cpu=0m on Node diek3dua9gay-3 + Feb 25 10:21:41.179: INFO: Pod cilium-2cncz requesting resource cpu=0m on Node diek3dua9gay-3 + Feb 25 10:21:41.179: INFO: Pod cilium-9tkfg requesting resource cpu=0m on Node diek3dua9gay-2 + Feb 25 10:21:41.179: INFO: Pod cilium-node-init-8b7n2 requesting resource cpu=100m on Node diek3dua9gay-2 + Feb 25 10:21:41.179: INFO: Pod cilium-node-init-979mb requesting resource cpu=100m on Node diek3dua9gay-3 + Feb 25 10:21:41.179: INFO: Pod cilium-node-init-s84z5 requesting resource cpu=100m on Node diek3dua9gay-1 + Feb 25 10:21:41.179: INFO: Pod cilium-operator-59d7d77cdc-lppbd requesting resource cpu=0m on Node diek3dua9gay-2 + Feb 25 10:21:41.179: INFO: Pod cilium-r4lc4 requesting resource cpu=0m on Node diek3dua9gay-1 + Feb 25 10:21:41.179: INFO: Pod coredns-565d847f94-5nh2n requesting resource cpu=100m on Node diek3dua9gay-2 + Feb 25 10:21:41.179: INFO: Pod coredns-565d847f94-qqvf6 requesting resource cpu=100m on Node diek3dua9gay-1 + Feb 25 10:21:41.179: INFO: Pod kube-addon-manager-diek3dua9gay-1 requesting resource cpu=5m on Node diek3dua9gay-1 + Feb 25 10:21:41.179: INFO: Pod kube-addon-manager-diek3dua9gay-2 requesting resource cpu=5m on Node diek3dua9gay-2 + Feb 25 10:21:41.179: INFO: Pod kube-apiserver-diek3dua9gay-1 requesting resource cpu=250m on Node diek3dua9gay-1 + Feb 25 10:21:41.179: INFO: Pod kube-apiserver-diek3dua9gay-2 requesting resource cpu=250m on Node diek3dua9gay-2 + Feb 25 10:21:41.179: INFO: Pod kube-controller-manager-diek3dua9gay-1 requesting resource cpu=200m on Node diek3dua9gay-1 + Feb 25 10:21:41.179: INFO: Pod kube-controller-manager-diek3dua9gay-2 requesting resource cpu=200m on Node diek3dua9gay-2 + Feb 25 10:21:41.179: INFO: Pod kube-proxy-22s9q requesting resource cpu=0m on Node diek3dua9gay-3 + Feb 25 10:21:41.179: INFO: Pod kube-proxy-cm4qg requesting resource cpu=0m on Node diek3dua9gay-1 + Feb 25 10:21:41.179: INFO: Pod kube-proxy-d66m7 requesting resource cpu=0m on Node diek3dua9gay-2 + Feb 25 10:21:41.179: INFO: Pod kube-scheduler-diek3dua9gay-1 requesting resource cpu=100m on Node diek3dua9gay-1 + Feb 25 10:21:41.179: INFO: Pod kube-scheduler-diek3dua9gay-2 requesting resource cpu=100m on Node diek3dua9gay-2 + Feb 25 10:21:41.179: INFO: Pod sonobuoy requesting resource cpu=0m on Node diek3dua9gay-3 + Feb 25 10:21:41.179: INFO: Pod sonobuoy-e2e-job-0b3c30975b5c4880 requesting resource cpu=0m on Node diek3dua9gay-3 + Feb 25 10:21:41.179: INFO: Pod sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-9pdjb requesting resource cpu=0m on Node diek3dua9gay-3 + Feb 25 10:21:41.179: INFO: Pod sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-mml8w requesting resource cpu=0m on Node diek3dua9gay-2 + Feb 25 10:21:41.179: INFO: Pod sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-wznrx requesting resource cpu=0m on Node diek3dua9gay-1 + STEP: Starting Pods to consume most of the cluster CPU. 02/25/23 10:21:41.179 + Feb 25 10:21:41.179: INFO: Creating a pod which consumes cpu=591m on Node diek3dua9gay-1 + Feb 25 10:21:41.460: INFO: Creating a pod which consumes cpu=591m on Node diek3dua9gay-2 + Feb 25 10:21:41.631: INFO: Creating a pod which consumes cpu=1050m on Node diek3dua9gay-3 + Feb 25 10:21:41.898: INFO: Waiting up to 5m0s for pod "filler-pod-b8a42890-7232-410e-bc1d-0832f167939b" in namespace "sched-pred-7978" to be "running" + Feb 25 10:21:42.361: INFO: Pod "filler-pod-b8a42890-7232-410e-bc1d-0832f167939b": Phase="Pending", Reason="", readiness=false. Elapsed: 462.663176ms + Feb 25 10:21:44.778: INFO: Pod "filler-pod-b8a42890-7232-410e-bc1d-0832f167939b": Phase="Running", Reason="", readiness=true. Elapsed: 2.879304518s + Feb 25 10:21:44.778: INFO: Pod "filler-pod-b8a42890-7232-410e-bc1d-0832f167939b" satisfied condition "running" + Feb 25 10:21:44.778: INFO: Waiting up to 5m0s for pod "filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a" in namespace "sched-pred-7978" to be "running" + Feb 25 10:21:44.823: INFO: Pod "filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a": Phase="Running", Reason="", readiness=true. Elapsed: 45.187474ms + Feb 25 10:21:45.069: INFO: Pod "filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a" satisfied condition "running" + Feb 25 10:21:45.069: INFO: Waiting up to 5m0s for pod "filler-pod-ecc20828-f391-4389-a241-36eb17240621" in namespace "sched-pred-7978" to be "running" + Feb 25 10:21:45.134: INFO: Pod "filler-pod-ecc20828-f391-4389-a241-36eb17240621": Phase="Pending", Reason="", readiness=false. Elapsed: 64.38347ms + Feb 25 10:21:47.144: INFO: Pod "filler-pod-ecc20828-f391-4389-a241-36eb17240621": Phase="Running", Reason="", readiness=true. Elapsed: 2.074499355s + Feb 25 10:21:47.144: INFO: Pod "filler-pod-ecc20828-f391-4389-a241-36eb17240621" satisfied condition "running" + STEP: Creating another pod that requires unavailable amount of CPU. 02/25/23 10:21:47.144 + STEP: Considering event: + Type = [Normal], Name = [filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a.17470a1c0353a83a], Reason = [Scheduled], Message = [Successfully assigned sched-pred-7978/filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a to diek3dua9gay-2] 02/25/23 10:21:47.153 + STEP: Considering event: + Type = [Normal], Name = [filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a.17470a1c5c70a2d1], Reason = [Pulled], Message = [Container image "registry.k8s.io/pause:3.8" already present on machine] 02/25/23 10:21:47.155 + STEP: Considering event: + Type = [Normal], Name = [filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a.17470a1c77b9d8a0], Reason = [Created], Message = [Created container filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a] 02/25/23 10:21:47.156 + STEP: Considering event: + Type = [Normal], Name = [filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a.17470a1c7a598e1c], Reason = [Started], Message = [Started container filler-pod-3a7ae03b-9cd7-4f19-8529-4c4c53a4483a] 02/25/23 10:21:47.156 + STEP: Considering event: + Type = [Normal], Name = [filler-pod-b8a42890-7232-410e-bc1d-0832f167939b.17470a1c012bb488], Reason = [Scheduled], Message = [Successfully assigned sched-pred-7978/filler-pod-b8a42890-7232-410e-bc1d-0832f167939b to diek3dua9gay-1] 02/25/23 10:21:47.156 + STEP: Considering event: + Type = [Normal], Name = [filler-pod-b8a42890-7232-410e-bc1d-0832f167939b.17470a1c75a411f8], Reason = [Pulled], Message = [Container image "registry.k8s.io/pause:3.8" already present on machine] 02/25/23 10:21:47.156 + STEP: Considering event: + Type = [Normal], Name = [filler-pod-b8a42890-7232-410e-bc1d-0832f167939b.17470a1c9195d22c], Reason = [Created], Message = [Created container filler-pod-b8a42890-7232-410e-bc1d-0832f167939b] 02/25/23 10:21:47.157 + STEP: Considering event: + Type = [Normal], Name = [filler-pod-b8a42890-7232-410e-bc1d-0832f167939b.17470a1c97cabc5b], Reason = [Started], Message = [Started container filler-pod-b8a42890-7232-410e-bc1d-0832f167939b] 02/25/23 10:21:47.157 + STEP: Considering event: + Type = [Normal], Name = [filler-pod-ecc20828-f391-4389-a241-36eb17240621.17470a1c0dad586a], Reason = [Scheduled], Message = [Successfully assigned sched-pred-7978/filler-pod-ecc20828-f391-4389-a241-36eb17240621 to diek3dua9gay-3] 02/25/23 10:21:47.158 + STEP: Considering event: + Type = [Normal], Name = [filler-pod-ecc20828-f391-4389-a241-36eb17240621.17470a1cdfc1a099], Reason = [Pulled], Message = [Container image "registry.k8s.io/pause:3.8" already present on machine] 02/25/23 10:21:47.158 + STEP: Considering event: + Type = [Normal], Name = [filler-pod-ecc20828-f391-4389-a241-36eb17240621.17470a1d0fe43b6b], Reason = [Created], Message = [Created container filler-pod-ecc20828-f391-4389-a241-36eb17240621] 02/25/23 10:21:47.158 + STEP: Considering event: + Type = [Normal], Name = [filler-pod-ecc20828-f391-4389-a241-36eb17240621.17470a1d1489bff1], Reason = [Started], Message = [Started container filler-pod-ecc20828-f391-4389-a241-36eb17240621] 02/25/23 10:21:47.158 + STEP: Considering event: + Type = [Warning], Name = [additional-pod.17470a1d4df2f4de], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 Insufficient cpu. preemption: 0/3 nodes are available: 3 No preemption victims found for incoming pod.] 02/25/23 10:21:47.202 + STEP: removing the label node off the node diek3dua9gay-1 02/25/23 10:21:48.181 + STEP: verifying the node doesn't have the label node 02/25/23 10:21:48.219 + STEP: removing the label node off the node diek3dua9gay-2 02/25/23 10:21:48.228 + STEP: verifying the node doesn't have the label node 02/25/23 10:21:48.252 + STEP: removing the label node off the node diek3dua9gay-3 02/25/23 10:21:48.262 + STEP: verifying the node doesn't have the label node 02/25/23 10:21:48.293 + [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:187 - Dec 14 16:22:54.825: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-8096" for this suite. 12/14/22 16:22:54.831 + Feb 25 10:21:48.317: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "sched-pred-7978" for this suite. 02/25/23 10:21:48.333 + [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:83 << End Captured GinkgoWriter Output ------------------------------ -[sig-storage] Downward API volume - should provide container's cpu limit [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:192 -[BeforeEach] [sig-storage] Downward API volume +SSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-api-machinery] Garbage collector + should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] + test/e2e/apimachinery/garbage_collector.go:735 +[BeforeEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:22:54.839 -Dec 14 16:22:54.839: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 16:22:54.842 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:54.863 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:54.866 -[BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 -[It] should provide container's cpu limit [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:192 -STEP: Creating a pod to test downward API volume plugin 12/14/22 16:22:54.869 -Dec 14 16:22:54.879: INFO: Waiting up to 5m0s for pod "downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9" in namespace "downward-api-9196" to be "Succeeded or Failed" -Dec 14 16:22:54.884: INFO: Pod "downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9": Phase="Pending", Reason="", readiness=false. Elapsed: 4.760403ms -Dec 14 16:22:56.893: INFO: Pod "downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013793995s -Dec 14 16:22:58.895: INFO: Pod "downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016034297s -STEP: Saw pod success 12/14/22 16:22:58.896 -Dec 14 16:22:58.896: INFO: Pod "downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9" satisfied condition "Succeeded or Failed" -Dec 14 16:22:58.902: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9 container client-container: -STEP: delete the pod 12/14/22 16:22:58.914 -Dec 14 16:22:58.932: INFO: Waiting for pod downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9 to disappear -Dec 14 16:22:58.936: INFO: Pod downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9 no longer exists -[AfterEach] [sig-storage] Downward API volume +STEP: Creating a kubernetes client 02/25/23 10:21:48.377 +Feb 25 10:21:48.378: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename gc 02/25/23 10:21:48.393 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:21:48.452 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:21:48.463 +[It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] + test/e2e/apimachinery/garbage_collector.go:735 +STEP: create the rc1 02/25/23 10:21:48.505 +STEP: create the rc2 02/25/23 10:21:48.545 +STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well 02/25/23 10:21:58.818 +STEP: delete the rc simpletest-rc-to-be-deleted 02/25/23 10:22:06.152 +STEP: wait for the rc to be deleted 02/25/23 10:22:06.247 +Feb 25 10:22:12.709: INFO: 94 pods remaining +Feb 25 10:22:12.709: INFO: 69 pods has nil DeletionTimestamp +Feb 25 10:22:12.709: INFO: +Feb 25 10:22:16.330: INFO: 82 pods remaining +Feb 25 10:22:16.330: INFO: 50 pods has nil DeletionTimestamp +Feb 25 10:22:16.330: INFO: +Feb 25 10:22:21.768: INFO: 51 pods remaining +Feb 25 10:22:21.768: INFO: 50 pods has nil DeletionTimestamp +Feb 25 10:22:21.768: INFO: +STEP: Gathering metrics 02/25/23 10:22:26.331 +Feb 25 10:22:26.650: INFO: Waiting up to 5m0s for pod "kube-controller-manager-diek3dua9gay-2" in namespace "kube-system" to be "running and ready" +Feb 25 10:22:26.660: INFO: Pod "kube-controller-manager-diek3dua9gay-2": Phase="Running", Reason="", readiness=true. Elapsed: 10.603523ms +Feb 25 10:22:26.661: INFO: The phase of Pod kube-controller-manager-diek3dua9gay-2 is Running (Ready = true) +Feb 25 10:22:26.661: INFO: Pod "kube-controller-manager-diek3dua9gay-2" satisfied condition "running and ready" +Feb 25 10:22:26.829: INFO: For apiserver_request_total: +For apiserver_request_latency_seconds: +For apiserver_init_events_total: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: + +Feb 25 10:22:26.829: INFO: Deleting pod "simpletest-rc-to-be-deleted-22smt" in namespace "gc-6591" +Feb 25 10:22:26.907: INFO: Deleting pod "simpletest-rc-to-be-deleted-2smlk" in namespace "gc-6591" +Feb 25 10:22:27.031: INFO: Deleting pod "simpletest-rc-to-be-deleted-4fz7q" in namespace "gc-6591" +Feb 25 10:22:27.080: INFO: Deleting pod "simpletest-rc-to-be-deleted-4lbz6" in namespace "gc-6591" +Feb 25 10:22:27.290: INFO: Deleting pod "simpletest-rc-to-be-deleted-4shzq" in namespace "gc-6591" +Feb 25 10:22:27.360: INFO: Deleting pod "simpletest-rc-to-be-deleted-4vwz2" in namespace "gc-6591" +Feb 25 10:22:27.633: INFO: Deleting pod "simpletest-rc-to-be-deleted-54lfv" in namespace "gc-6591" +Feb 25 10:22:27.657: INFO: Deleting pod "simpletest-rc-to-be-deleted-5kmkf" in namespace "gc-6591" +Feb 25 10:22:27.740: INFO: Deleting pod "simpletest-rc-to-be-deleted-5kv76" in namespace "gc-6591" +Feb 25 10:22:27.824: INFO: Deleting pod "simpletest-rc-to-be-deleted-5vndw" in namespace "gc-6591" +Feb 25 10:22:27.881: INFO: Deleting pod "simpletest-rc-to-be-deleted-6bq48" in namespace "gc-6591" +Feb 25 10:22:27.957: INFO: Deleting pod "simpletest-rc-to-be-deleted-6kcjs" in namespace "gc-6591" +Feb 25 10:22:28.062: INFO: Deleting pod "simpletest-rc-to-be-deleted-6qv9z" in namespace "gc-6591" +Feb 25 10:22:28.133: INFO: Deleting pod "simpletest-rc-to-be-deleted-6vrvr" in namespace "gc-6591" +Feb 25 10:22:28.186: INFO: Deleting pod "simpletest-rc-to-be-deleted-6wg8k" in namespace "gc-6591" +Feb 25 10:22:28.309: INFO: Deleting pod "simpletest-rc-to-be-deleted-727j4" in namespace "gc-6591" +Feb 25 10:22:28.364: INFO: Deleting pod "simpletest-rc-to-be-deleted-72k4r" in namespace "gc-6591" +Feb 25 10:22:28.425: INFO: Deleting pod "simpletest-rc-to-be-deleted-74rkg" in namespace "gc-6591" +Feb 25 10:22:28.491: INFO: Deleting pod "simpletest-rc-to-be-deleted-7csnv" in namespace "gc-6591" +Feb 25 10:22:28.616: INFO: Deleting pod "simpletest-rc-to-be-deleted-85pq7" in namespace "gc-6591" +Feb 25 10:22:28.744: INFO: Deleting pod "simpletest-rc-to-be-deleted-8l4s4" in namespace "gc-6591" +Feb 25 10:22:28.883: INFO: Deleting pod "simpletest-rc-to-be-deleted-97gkn" in namespace "gc-6591" +Feb 25 10:22:29.669: INFO: Deleting pod "simpletest-rc-to-be-deleted-97mpn" in namespace "gc-6591" +Feb 25 10:22:29.848: INFO: Deleting pod "simpletest-rc-to-be-deleted-9fpvm" in namespace "gc-6591" +Feb 25 10:22:29.905: INFO: Deleting pod "simpletest-rc-to-be-deleted-9fsjw" in namespace "gc-6591" +Feb 25 10:22:30.018: INFO: Deleting pod "simpletest-rc-to-be-deleted-9gb2k" in namespace "gc-6591" +Feb 25 10:22:30.101: INFO: Deleting pod "simpletest-rc-to-be-deleted-9hsjg" in namespace "gc-6591" +Feb 25 10:22:30.185: INFO: Deleting pod "simpletest-rc-to-be-deleted-9lx9k" in namespace "gc-6591" +Feb 25 10:22:30.283: INFO: Deleting pod "simpletest-rc-to-be-deleted-9pcmn" in namespace "gc-6591" +Feb 25 10:22:30.325: INFO: Deleting pod "simpletest-rc-to-be-deleted-9qg8t" in namespace "gc-6591" +Feb 25 10:22:30.466: INFO: Deleting pod "simpletest-rc-to-be-deleted-9s2x5" in namespace "gc-6591" +Feb 25 10:22:30.570: INFO: Deleting pod "simpletest-rc-to-be-deleted-9xk6x" in namespace "gc-6591" +Feb 25 10:22:30.639: INFO: Deleting pod "simpletest-rc-to-be-deleted-b6tpp" in namespace "gc-6591" +Feb 25 10:22:30.744: INFO: Deleting pod "simpletest-rc-to-be-deleted-cbmhj" in namespace "gc-6591" +Feb 25 10:22:30.799: INFO: Deleting pod "simpletest-rc-to-be-deleted-cbs25" in namespace "gc-6591" +Feb 25 10:22:30.903: INFO: Deleting pod "simpletest-rc-to-be-deleted-cjd6b" in namespace "gc-6591" +Feb 25 10:22:31.097: INFO: Deleting pod "simpletest-rc-to-be-deleted-czzm5" in namespace "gc-6591" +Feb 25 10:22:31.213: INFO: Deleting pod "simpletest-rc-to-be-deleted-dxwrk" in namespace "gc-6591" +Feb 25 10:22:31.302: INFO: Deleting pod "simpletest-rc-to-be-deleted-fc4jd" in namespace "gc-6591" +Feb 25 10:22:31.442: INFO: Deleting pod "simpletest-rc-to-be-deleted-fdqmd" in namespace "gc-6591" +Feb 25 10:22:31.648: INFO: Deleting pod "simpletest-rc-to-be-deleted-fmdcv" in namespace "gc-6591" +Feb 25 10:22:31.747: INFO: Deleting pod "simpletest-rc-to-be-deleted-g2c7l" in namespace "gc-6591" +Feb 25 10:22:31.863: INFO: Deleting pod "simpletest-rc-to-be-deleted-g9np2" in namespace "gc-6591" +Feb 25 10:22:31.959: INFO: Deleting pod "simpletest-rc-to-be-deleted-gb2ph" in namespace "gc-6591" +Feb 25 10:22:32.028: INFO: Deleting pod "simpletest-rc-to-be-deleted-gg5bd" in namespace "gc-6591" +Feb 25 10:22:32.106: INFO: Deleting pod "simpletest-rc-to-be-deleted-gnt5h" in namespace "gc-6591" +Feb 25 10:22:32.172: INFO: Deleting pod "simpletest-rc-to-be-deleted-gsn4g" in namespace "gc-6591" +Feb 25 10:22:32.225: INFO: Deleting pod "simpletest-rc-to-be-deleted-gvhqt" in namespace "gc-6591" +Feb 25 10:22:32.364: INFO: Deleting pod "simpletest-rc-to-be-deleted-h46s5" in namespace "gc-6591" +Feb 25 10:22:32.466: INFO: Deleting pod "simpletest-rc-to-be-deleted-h9xs6" in namespace "gc-6591" +[AfterEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:187 -Dec 14 16:22:58.936: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-9196" for this suite. 12/14/22 16:22:58.942 -{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu limit [NodeConformance] [Conformance]","completed":133,"skipped":2378,"failed":0} +Feb 25 10:22:32.555: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "gc-6591" for this suite. 02/25/23 10:22:32.571 +{"msg":"PASSED [sig-api-machinery] Garbage collector should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance]","completed":127,"skipped":2368,"failed":0} ------------------------------ -• [4.113 seconds] -[sig-storage] Downward API volume -test/e2e/common/storage/framework.go:23 - should provide container's cpu limit [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:192 +• [SLOW TEST] [44.238 seconds] +[sig-api-machinery] Garbage collector +test/e2e/apimachinery/framework.go:23 + should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] + test/e2e/apimachinery/garbage_collector.go:735 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-api-machinery] Garbage collector + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:21:48.377 + Feb 25 10:21:48.378: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename gc 02/25/23 10:21:48.393 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:21:48.452 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:21:48.463 + [It] should not delete dependents that have both valid owner and owner that's waiting for dependents to be deleted [Conformance] + test/e2e/apimachinery/garbage_collector.go:735 + STEP: create the rc1 02/25/23 10:21:48.505 + STEP: create the rc2 02/25/23 10:21:48.545 + STEP: set half of pods created by rc simpletest-rc-to-be-deleted to have rc simpletest-rc-to-stay as owner as well 02/25/23 10:21:58.818 + STEP: delete the rc simpletest-rc-to-be-deleted 02/25/23 10:22:06.152 + STEP: wait for the rc to be deleted 02/25/23 10:22:06.247 + Feb 25 10:22:12.709: INFO: 94 pods remaining + Feb 25 10:22:12.709: INFO: 69 pods has nil DeletionTimestamp + Feb 25 10:22:12.709: INFO: + Feb 25 10:22:16.330: INFO: 82 pods remaining + Feb 25 10:22:16.330: INFO: 50 pods has nil DeletionTimestamp + Feb 25 10:22:16.330: INFO: + Feb 25 10:22:21.768: INFO: 51 pods remaining + Feb 25 10:22:21.768: INFO: 50 pods has nil DeletionTimestamp + Feb 25 10:22:21.768: INFO: + STEP: Gathering metrics 02/25/23 10:22:26.331 + Feb 25 10:22:26.650: INFO: Waiting up to 5m0s for pod "kube-controller-manager-diek3dua9gay-2" in namespace "kube-system" to be "running and ready" + Feb 25 10:22:26.660: INFO: Pod "kube-controller-manager-diek3dua9gay-2": Phase="Running", Reason="", readiness=true. Elapsed: 10.603523ms + Feb 25 10:22:26.661: INFO: The phase of Pod kube-controller-manager-diek3dua9gay-2 is Running (Ready = true) + Feb 25 10:22:26.661: INFO: Pod "kube-controller-manager-diek3dua9gay-2" satisfied condition "running and ready" + Feb 25 10:22:26.829: INFO: For apiserver_request_total: + For apiserver_request_latency_seconds: + For apiserver_init_events_total: + For garbage_collector_attempt_to_delete_queue_latency: + For garbage_collector_attempt_to_delete_work_duration: + For garbage_collector_attempt_to_orphan_queue_latency: + For garbage_collector_attempt_to_orphan_work_duration: + For garbage_collector_dirty_processing_latency_microseconds: + For garbage_collector_event_processing_latency_microseconds: + For garbage_collector_graph_changes_queue_latency: + For garbage_collector_graph_changes_work_duration: + For garbage_collector_orphan_processing_latency_microseconds: + For namespace_queue_latency: + For namespace_queue_latency_sum: + For namespace_queue_latency_count: + For namespace_retries: + For namespace_work_duration: + For namespace_work_duration_sum: + For namespace_work_duration_count: + For function_duration_seconds: + For errors_total: + For evicted_pods_total: - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Downward API volume - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:22:54.839 - Dec 14 16:22:54.839: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 16:22:54.842 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:54.863 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:54.866 - [BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 - [It] should provide container's cpu limit [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:192 - STEP: Creating a pod to test downward API volume plugin 12/14/22 16:22:54.869 - Dec 14 16:22:54.879: INFO: Waiting up to 5m0s for pod "downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9" in namespace "downward-api-9196" to be "Succeeded or Failed" - Dec 14 16:22:54.884: INFO: Pod "downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9": Phase="Pending", Reason="", readiness=false. Elapsed: 4.760403ms - Dec 14 16:22:56.893: INFO: Pod "downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013793995s - Dec 14 16:22:58.895: INFO: Pod "downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016034297s - STEP: Saw pod success 12/14/22 16:22:58.896 - Dec 14 16:22:58.896: INFO: Pod "downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9" satisfied condition "Succeeded or Failed" - Dec 14 16:22:58.902: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9 container client-container: - STEP: delete the pod 12/14/22 16:22:58.914 - Dec 14 16:22:58.932: INFO: Waiting for pod downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9 to disappear - Dec 14 16:22:58.936: INFO: Pod downwardapi-volume-3a44c498-e1ed-4f7e-9b64-93a42af11be9 no longer exists - [AfterEach] [sig-storage] Downward API volume + Feb 25 10:22:26.829: INFO: Deleting pod "simpletest-rc-to-be-deleted-22smt" in namespace "gc-6591" + Feb 25 10:22:26.907: INFO: Deleting pod "simpletest-rc-to-be-deleted-2smlk" in namespace "gc-6591" + Feb 25 10:22:27.031: INFO: Deleting pod "simpletest-rc-to-be-deleted-4fz7q" in namespace "gc-6591" + Feb 25 10:22:27.080: INFO: Deleting pod "simpletest-rc-to-be-deleted-4lbz6" in namespace "gc-6591" + Feb 25 10:22:27.290: INFO: Deleting pod "simpletest-rc-to-be-deleted-4shzq" in namespace "gc-6591" + Feb 25 10:22:27.360: INFO: Deleting pod "simpletest-rc-to-be-deleted-4vwz2" in namespace "gc-6591" + Feb 25 10:22:27.633: INFO: Deleting pod "simpletest-rc-to-be-deleted-54lfv" in namespace "gc-6591" + Feb 25 10:22:27.657: INFO: Deleting pod "simpletest-rc-to-be-deleted-5kmkf" in namespace "gc-6591" + Feb 25 10:22:27.740: INFO: Deleting pod "simpletest-rc-to-be-deleted-5kv76" in namespace "gc-6591" + Feb 25 10:22:27.824: INFO: Deleting pod "simpletest-rc-to-be-deleted-5vndw" in namespace "gc-6591" + Feb 25 10:22:27.881: INFO: Deleting pod "simpletest-rc-to-be-deleted-6bq48" in namespace "gc-6591" + Feb 25 10:22:27.957: INFO: Deleting pod "simpletest-rc-to-be-deleted-6kcjs" in namespace "gc-6591" + Feb 25 10:22:28.062: INFO: Deleting pod "simpletest-rc-to-be-deleted-6qv9z" in namespace "gc-6591" + Feb 25 10:22:28.133: INFO: Deleting pod "simpletest-rc-to-be-deleted-6vrvr" in namespace "gc-6591" + Feb 25 10:22:28.186: INFO: Deleting pod "simpletest-rc-to-be-deleted-6wg8k" in namespace "gc-6591" + Feb 25 10:22:28.309: INFO: Deleting pod "simpletest-rc-to-be-deleted-727j4" in namespace "gc-6591" + Feb 25 10:22:28.364: INFO: Deleting pod "simpletest-rc-to-be-deleted-72k4r" in namespace "gc-6591" + Feb 25 10:22:28.425: INFO: Deleting pod "simpletest-rc-to-be-deleted-74rkg" in namespace "gc-6591" + Feb 25 10:22:28.491: INFO: Deleting pod "simpletest-rc-to-be-deleted-7csnv" in namespace "gc-6591" + Feb 25 10:22:28.616: INFO: Deleting pod "simpletest-rc-to-be-deleted-85pq7" in namespace "gc-6591" + Feb 25 10:22:28.744: INFO: Deleting pod "simpletest-rc-to-be-deleted-8l4s4" in namespace "gc-6591" + Feb 25 10:22:28.883: INFO: Deleting pod "simpletest-rc-to-be-deleted-97gkn" in namespace "gc-6591" + Feb 25 10:22:29.669: INFO: Deleting pod "simpletest-rc-to-be-deleted-97mpn" in namespace "gc-6591" + Feb 25 10:22:29.848: INFO: Deleting pod "simpletest-rc-to-be-deleted-9fpvm" in namespace "gc-6591" + Feb 25 10:22:29.905: INFO: Deleting pod "simpletest-rc-to-be-deleted-9fsjw" in namespace "gc-6591" + Feb 25 10:22:30.018: INFO: Deleting pod "simpletest-rc-to-be-deleted-9gb2k" in namespace "gc-6591" + Feb 25 10:22:30.101: INFO: Deleting pod "simpletest-rc-to-be-deleted-9hsjg" in namespace "gc-6591" + Feb 25 10:22:30.185: INFO: Deleting pod "simpletest-rc-to-be-deleted-9lx9k" in namespace "gc-6591" + Feb 25 10:22:30.283: INFO: Deleting pod "simpletest-rc-to-be-deleted-9pcmn" in namespace "gc-6591" + Feb 25 10:22:30.325: INFO: Deleting pod "simpletest-rc-to-be-deleted-9qg8t" in namespace "gc-6591" + Feb 25 10:22:30.466: INFO: Deleting pod "simpletest-rc-to-be-deleted-9s2x5" in namespace "gc-6591" + Feb 25 10:22:30.570: INFO: Deleting pod "simpletest-rc-to-be-deleted-9xk6x" in namespace "gc-6591" + Feb 25 10:22:30.639: INFO: Deleting pod "simpletest-rc-to-be-deleted-b6tpp" in namespace "gc-6591" + Feb 25 10:22:30.744: INFO: Deleting pod "simpletest-rc-to-be-deleted-cbmhj" in namespace "gc-6591" + Feb 25 10:22:30.799: INFO: Deleting pod "simpletest-rc-to-be-deleted-cbs25" in namespace "gc-6591" + Feb 25 10:22:30.903: INFO: Deleting pod "simpletest-rc-to-be-deleted-cjd6b" in namespace "gc-6591" + Feb 25 10:22:31.097: INFO: Deleting pod "simpletest-rc-to-be-deleted-czzm5" in namespace "gc-6591" + Feb 25 10:22:31.213: INFO: Deleting pod "simpletest-rc-to-be-deleted-dxwrk" in namespace "gc-6591" + Feb 25 10:22:31.302: INFO: Deleting pod "simpletest-rc-to-be-deleted-fc4jd" in namespace "gc-6591" + Feb 25 10:22:31.442: INFO: Deleting pod "simpletest-rc-to-be-deleted-fdqmd" in namespace "gc-6591" + Feb 25 10:22:31.648: INFO: Deleting pod "simpletest-rc-to-be-deleted-fmdcv" in namespace "gc-6591" + Feb 25 10:22:31.747: INFO: Deleting pod "simpletest-rc-to-be-deleted-g2c7l" in namespace "gc-6591" + Feb 25 10:22:31.863: INFO: Deleting pod "simpletest-rc-to-be-deleted-g9np2" in namespace "gc-6591" + Feb 25 10:22:31.959: INFO: Deleting pod "simpletest-rc-to-be-deleted-gb2ph" in namespace "gc-6591" + Feb 25 10:22:32.028: INFO: Deleting pod "simpletest-rc-to-be-deleted-gg5bd" in namespace "gc-6591" + Feb 25 10:22:32.106: INFO: Deleting pod "simpletest-rc-to-be-deleted-gnt5h" in namespace "gc-6591" + Feb 25 10:22:32.172: INFO: Deleting pod "simpletest-rc-to-be-deleted-gsn4g" in namespace "gc-6591" + Feb 25 10:22:32.225: INFO: Deleting pod "simpletest-rc-to-be-deleted-gvhqt" in namespace "gc-6591" + Feb 25 10:22:32.364: INFO: Deleting pod "simpletest-rc-to-be-deleted-h46s5" in namespace "gc-6591" + Feb 25 10:22:32.466: INFO: Deleting pod "simpletest-rc-to-be-deleted-h9xs6" in namespace "gc-6591" + [AfterEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:187 - Dec 14 16:22:58.936: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-9196" for this suite. 12/14/22 16:22:58.942 + Feb 25 10:22:32.555: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "gc-6591" for this suite. 02/25/23 10:22:32.571 << End Captured GinkgoWriter Output ------------------------------ -S +SSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-network] Proxy version v1 - A set of valid responses are returned for both pod and service ProxyWithPath [Conformance] - test/e2e/network/proxy.go:286 -[BeforeEach] version v1 +[sig-api-machinery] ResourceQuota + should create a ResourceQuota and capture the life of a pod. [Conformance] + test/e2e/apimachinery/resource_quota.go:220 +[BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:22:58.958 -Dec 14 16:22:58.958: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename proxy 12/14/22 16:22:58.962 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:59 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:59.011 -[It] A set of valid responses are returned for both pod and service ProxyWithPath [Conformance] - test/e2e/network/proxy.go:286 -Dec 14 16:22:59.016: INFO: Creating pod... -Dec 14 16:22:59.042: INFO: Waiting up to 5m0s for pod "agnhost" in namespace "proxy-5264" to be "running" -Dec 14 16:22:59.058: INFO: Pod "agnhost": Phase="Pending", Reason="", readiness=false. Elapsed: 16.737917ms -Dec 14 16:23:01.064: INFO: Pod "agnhost": Phase="Running", Reason="", readiness=true. Elapsed: 2.022922391s -Dec 14 16:23:01.065: INFO: Pod "agnhost" satisfied condition "running" -Dec 14 16:23:01.065: INFO: Creating service... -Dec 14 16:23:01.079: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/pods/agnhost/proxy/some/path/with/DELETE -Dec 14 16:23:01.132: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE -Dec 14 16:23:01.133: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/pods/agnhost/proxy/some/path/with/GET -Dec 14 16:23:01.145: INFO: http.Client request:GET | StatusCode:200 | Response:foo | Method:GET -Dec 14 16:23:01.145: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/pods/agnhost/proxy/some/path/with/HEAD -Dec 14 16:23:01.164: INFO: http.Client request:HEAD | StatusCode:200 -Dec 14 16:23:01.164: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/pods/agnhost/proxy/some/path/with/OPTIONS -Dec 14 16:23:01.170: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS -Dec 14 16:23:01.170: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/pods/agnhost/proxy/some/path/with/PATCH -Dec 14 16:23:01.176: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH -Dec 14 16:23:01.176: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/pods/agnhost/proxy/some/path/with/POST -Dec 14 16:23:01.180: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST -Dec 14 16:23:01.180: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/pods/agnhost/proxy/some/path/with/PUT -Dec 14 16:23:01.186: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT -Dec 14 16:23:01.186: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/services/test-service/proxy/some/path/with/DELETE -Dec 14 16:23:01.194: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE -Dec 14 16:23:01.194: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/services/test-service/proxy/some/path/with/GET -Dec 14 16:23:01.200: INFO: http.Client request:GET | StatusCode:200 | Response:foo | Method:GET -Dec 14 16:23:01.200: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/services/test-service/proxy/some/path/with/HEAD -Dec 14 16:23:01.206: INFO: http.Client request:HEAD | StatusCode:200 -Dec 14 16:23:01.206: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/services/test-service/proxy/some/path/with/OPTIONS -Dec 14 16:23:01.213: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS -Dec 14 16:23:01.213: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/services/test-service/proxy/some/path/with/PATCH -Dec 14 16:23:01.219: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH -Dec 14 16:23:01.219: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/services/test-service/proxy/some/path/with/POST -Dec 14 16:23:01.228: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST -Dec 14 16:23:01.228: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/services/test-service/proxy/some/path/with/PUT -Dec 14 16:23:01.260: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT -[AfterEach] version v1 +STEP: Creating a kubernetes client 02/25/23 10:22:32.653 +Feb 25 10:22:32.653: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename resourcequota 02/25/23 10:22:32.658 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:22:32.727 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:22:32.734 +[It] should create a ResourceQuota and capture the life of a pod. [Conformance] + test/e2e/apimachinery/resource_quota.go:220 +STEP: Counting existing ResourceQuota 02/25/23 10:22:32.743 +STEP: Creating a ResourceQuota 02/25/23 10:22:37.752 +STEP: Ensuring resource quota status is calculated 02/25/23 10:22:37.777 +STEP: Creating a Pod that fits quota 02/25/23 10:22:39.808 +STEP: Ensuring ResourceQuota status captures the pod usage 02/25/23 10:22:39.926 +STEP: Not allowing a pod to be created that exceeds remaining quota 02/25/23 10:22:41.937 +STEP: Not allowing a pod to be created that exceeds remaining quota(validation on extended resources) 02/25/23 10:22:41.943 +STEP: Ensuring a pod cannot update its resource requirements 02/25/23 10:22:41.947 +STEP: Ensuring attempts to update pod resource requirements did not change quota usage 02/25/23 10:22:41.959 +STEP: Deleting the pod 02/25/23 10:22:43.973 +STEP: Ensuring resource quota status released the pod usage 02/25/23 10:22:44.008 +[AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 -Dec 14 16:23:01.261: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "proxy-5264" for this suite. 12/14/22 16:23:01.275 -{"msg":"PASSED [sig-network] Proxy version v1 A set of valid responses are returned for both pod and service ProxyWithPath [Conformance]","completed":134,"skipped":2379,"failed":0} +Feb 25 10:22:46.018: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-3654" for this suite. 02/25/23 10:22:46.053 +{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a pod. [Conformance]","completed":128,"skipped":2394,"failed":0} ------------------------------ -• [2.327 seconds] -[sig-network] Proxy -test/e2e/network/common/framework.go:23 - version v1 - test/e2e/network/proxy.go:74 - A set of valid responses are returned for both pod and service ProxyWithPath [Conformance] - test/e2e/network/proxy.go:286 +• [SLOW TEST] [13.422 seconds] +[sig-api-machinery] ResourceQuota +test/e2e/apimachinery/framework.go:23 + should create a ResourceQuota and capture the life of a pod. [Conformance] + test/e2e/apimachinery/resource_quota.go:220 Begin Captured GinkgoWriter Output >> - [BeforeEach] version v1 + [BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:22:58.958 - Dec 14 16:22:58.958: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename proxy 12/14/22 16:22:58.962 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:22:59 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:22:59.011 - [It] A set of valid responses are returned for both pod and service ProxyWithPath [Conformance] - test/e2e/network/proxy.go:286 - Dec 14 16:22:59.016: INFO: Creating pod... - Dec 14 16:22:59.042: INFO: Waiting up to 5m0s for pod "agnhost" in namespace "proxy-5264" to be "running" - Dec 14 16:22:59.058: INFO: Pod "agnhost": Phase="Pending", Reason="", readiness=false. Elapsed: 16.737917ms - Dec 14 16:23:01.064: INFO: Pod "agnhost": Phase="Running", Reason="", readiness=true. Elapsed: 2.022922391s - Dec 14 16:23:01.065: INFO: Pod "agnhost" satisfied condition "running" - Dec 14 16:23:01.065: INFO: Creating service... - Dec 14 16:23:01.079: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/pods/agnhost/proxy/some/path/with/DELETE - Dec 14 16:23:01.132: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE - Dec 14 16:23:01.133: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/pods/agnhost/proxy/some/path/with/GET - Dec 14 16:23:01.145: INFO: http.Client request:GET | StatusCode:200 | Response:foo | Method:GET - Dec 14 16:23:01.145: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/pods/agnhost/proxy/some/path/with/HEAD - Dec 14 16:23:01.164: INFO: http.Client request:HEAD | StatusCode:200 - Dec 14 16:23:01.164: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/pods/agnhost/proxy/some/path/with/OPTIONS - Dec 14 16:23:01.170: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS - Dec 14 16:23:01.170: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/pods/agnhost/proxy/some/path/with/PATCH - Dec 14 16:23:01.176: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH - Dec 14 16:23:01.176: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/pods/agnhost/proxy/some/path/with/POST - Dec 14 16:23:01.180: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST - Dec 14 16:23:01.180: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/pods/agnhost/proxy/some/path/with/PUT - Dec 14 16:23:01.186: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT - Dec 14 16:23:01.186: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/services/test-service/proxy/some/path/with/DELETE - Dec 14 16:23:01.194: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE - Dec 14 16:23:01.194: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/services/test-service/proxy/some/path/with/GET - Dec 14 16:23:01.200: INFO: http.Client request:GET | StatusCode:200 | Response:foo | Method:GET - Dec 14 16:23:01.200: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/services/test-service/proxy/some/path/with/HEAD - Dec 14 16:23:01.206: INFO: http.Client request:HEAD | StatusCode:200 - Dec 14 16:23:01.206: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/services/test-service/proxy/some/path/with/OPTIONS - Dec 14 16:23:01.213: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS - Dec 14 16:23:01.213: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/services/test-service/proxy/some/path/with/PATCH - Dec 14 16:23:01.219: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH - Dec 14 16:23:01.219: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/services/test-service/proxy/some/path/with/POST - Dec 14 16:23:01.228: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST - Dec 14 16:23:01.228: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5264/services/test-service/proxy/some/path/with/PUT - Dec 14 16:23:01.260: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT - [AfterEach] version v1 + STEP: Creating a kubernetes client 02/25/23 10:22:32.653 + Feb 25 10:22:32.653: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename resourcequota 02/25/23 10:22:32.658 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:22:32.727 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:22:32.734 + [It] should create a ResourceQuota and capture the life of a pod. [Conformance] + test/e2e/apimachinery/resource_quota.go:220 + STEP: Counting existing ResourceQuota 02/25/23 10:22:32.743 + STEP: Creating a ResourceQuota 02/25/23 10:22:37.752 + STEP: Ensuring resource quota status is calculated 02/25/23 10:22:37.777 + STEP: Creating a Pod that fits quota 02/25/23 10:22:39.808 + STEP: Ensuring ResourceQuota status captures the pod usage 02/25/23 10:22:39.926 + STEP: Not allowing a pod to be created that exceeds remaining quota 02/25/23 10:22:41.937 + STEP: Not allowing a pod to be created that exceeds remaining quota(validation on extended resources) 02/25/23 10:22:41.943 + STEP: Ensuring a pod cannot update its resource requirements 02/25/23 10:22:41.947 + STEP: Ensuring attempts to update pod resource requirements did not change quota usage 02/25/23 10:22:41.959 + STEP: Deleting the pod 02/25/23 10:22:43.973 + STEP: Ensuring resource quota status released the pod usage 02/25/23 10:22:44.008 + [AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 - Dec 14 16:23:01.261: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "proxy-5264" for this suite. 12/14/22 16:23:01.275 + Feb 25 10:22:46.018: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "resourcequota-3654" for this suite. 02/25/23 10:22:46.053 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-apps] CronJob - should replace jobs when ReplaceConcurrent [Conformance] - test/e2e/apps/cronjob.go:160 -[BeforeEach] [sig-apps] CronJob +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + works for multiple CRDs of different groups [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:275 +[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:23:01.291 -Dec 14 16:23:01.291: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename cronjob 12/14/22 16:23:01.292 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:23:01.316 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:23:01.32 -[It] should replace jobs when ReplaceConcurrent [Conformance] - test/e2e/apps/cronjob.go:160 -STEP: Creating a ReplaceConcurrent cronjob 12/14/22 16:23:01.324 -STEP: Ensuring a job is scheduled 12/14/22 16:23:01.332 -STEP: Ensuring exactly one is scheduled 12/14/22 16:24:01.341 -STEP: Ensuring exactly one running job exists by listing jobs explicitly 12/14/22 16:24:01.347 -STEP: Ensuring the job is replaced with a new one 12/14/22 16:24:01.355 -STEP: Removing cronjob 12/14/22 16:25:01.365 -[AfterEach] [sig-apps] CronJob +STEP: Creating a kubernetes client 02/25/23 10:22:46.081 +Feb 25 10:22:46.081: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 10:22:46.086 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:22:46.144 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:22:46.152 +[It] works for multiple CRDs of different groups [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:275 +STEP: CRs in different groups (two CRDs) show up in OpenAPI documentation 02/25/23 10:22:46.158 +Feb 25 10:22:46.160: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:22:53.920: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:25:01.377: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "cronjob-6158" for this suite. 12/14/22 16:25:01.387 -{"msg":"PASSED [sig-apps] CronJob should replace jobs when ReplaceConcurrent [Conformance]","completed":135,"skipped":2421,"failed":0} +Feb 25 10:23:21.789: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "crd-publish-openapi-3604" for this suite. 02/25/23 10:23:21.808 +{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of different groups [Conformance]","completed":129,"skipped":2426,"failed":0} ------------------------------ -• [SLOW TEST] [120.109 seconds] -[sig-apps] CronJob -test/e2e/apps/framework.go:23 - should replace jobs when ReplaceConcurrent [Conformance] - test/e2e/apps/cronjob.go:160 +• [SLOW TEST] [35.742 seconds] +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + works for multiple CRDs of different groups [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:275 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] CronJob + [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:23:01.291 - Dec 14 16:23:01.291: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename cronjob 12/14/22 16:23:01.292 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:23:01.316 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:23:01.32 - [It] should replace jobs when ReplaceConcurrent [Conformance] - test/e2e/apps/cronjob.go:160 - STEP: Creating a ReplaceConcurrent cronjob 12/14/22 16:23:01.324 - STEP: Ensuring a job is scheduled 12/14/22 16:23:01.332 - STEP: Ensuring exactly one is scheduled 12/14/22 16:24:01.341 - STEP: Ensuring exactly one running job exists by listing jobs explicitly 12/14/22 16:24:01.347 - STEP: Ensuring the job is replaced with a new one 12/14/22 16:24:01.355 - STEP: Removing cronjob 12/14/22 16:25:01.365 - [AfterEach] [sig-apps] CronJob + STEP: Creating a kubernetes client 02/25/23 10:22:46.081 + Feb 25 10:22:46.081: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 10:22:46.086 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:22:46.144 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:22:46.152 + [It] works for multiple CRDs of different groups [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:275 + STEP: CRs in different groups (two CRDs) show up in OpenAPI documentation 02/25/23 10:22:46.158 + Feb 25 10:22:46.160: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:22:53.920: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:25:01.377: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "cronjob-6158" for this suite. 12/14/22 16:25:01.387 + Feb 25 10:23:21.789: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "crd-publish-openapi-3604" for this suite. 02/25/23 10:23:21.808 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSS +SSSSS ------------------------------ -[sig-scheduling] SchedulerPreemption [Serial] - validates lower priority pod preemption by critical pod [Conformance] - test/e2e/scheduling/preemption.go:218 -[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] +[sig-node] Security Context + should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] + test/e2e/node/security_context.go:132 +[BeforeEach] [sig-node] Security Context test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:25:01.402 -Dec 14 16:25:01.402: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename sched-preemption 12/14/22 16:25:01.406 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:25:01.46 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:25:01.466 -[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:92 -Dec 14 16:25:01.494: INFO: Waiting up to 1m0s for all nodes to be ready -Dec 14 16:26:01.535: INFO: Waiting for terminating namespaces to be deleted... -[It] validates lower priority pod preemption by critical pod [Conformance] - test/e2e/scheduling/preemption.go:218 -STEP: Create pods that use 4/5 of node resources. 12/14/22 16:26:01.541 -Dec 14 16:26:01.580: INFO: Created pod: pod0-0-sched-preemption-low-priority -Dec 14 16:26:01.589: INFO: Created pod: pod0-1-sched-preemption-medium-priority -Dec 14 16:26:01.614: INFO: Created pod: pod1-0-sched-preemption-medium-priority -Dec 14 16:26:01.621: INFO: Created pod: pod1-1-sched-preemption-medium-priority -Dec 14 16:26:01.648: INFO: Created pod: pod2-0-sched-preemption-medium-priority -Dec 14 16:26:01.658: INFO: Created pod: pod2-1-sched-preemption-medium-priority -STEP: Wait for pods to be scheduled. 12/14/22 16:26:01.658 -Dec 14 16:26:01.659: INFO: Waiting up to 5m0s for pod "pod0-0-sched-preemption-low-priority" in namespace "sched-preemption-8428" to be "running" -Dec 14 16:26:01.664: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 5.237849ms -Dec 14 16:26:03.674: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014987854s -Dec 14 16:26:05.674: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 4.01528966s -Dec 14 16:26:07.676: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Running", Reason="", readiness=true. Elapsed: 6.017220372s -Dec 14 16:26:07.676: INFO: Pod "pod0-0-sched-preemption-low-priority" satisfied condition "running" -Dec 14 16:26:07.676: INFO: Waiting up to 5m0s for pod "pod0-1-sched-preemption-medium-priority" in namespace "sched-preemption-8428" to be "running" -Dec 14 16:26:07.683: INFO: Pod "pod0-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 6.628224ms -Dec 14 16:26:07.683: INFO: Pod "pod0-1-sched-preemption-medium-priority" satisfied condition "running" -Dec 14 16:26:07.683: INFO: Waiting up to 5m0s for pod "pod1-0-sched-preemption-medium-priority" in namespace "sched-preemption-8428" to be "running" -Dec 14 16:26:07.690: INFO: Pod "pod1-0-sched-preemption-medium-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 6.501223ms -Dec 14 16:26:09.700: INFO: Pod "pod1-0-sched-preemption-medium-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017010901s -Dec 14 16:26:11.699: INFO: Pod "pod1-0-sched-preemption-medium-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 4.016155265s -Dec 14 16:26:13.701: INFO: Pod "pod1-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 6.017900562s -Dec 14 16:26:13.701: INFO: Pod "pod1-0-sched-preemption-medium-priority" satisfied condition "running" -Dec 14 16:26:13.701: INFO: Waiting up to 5m0s for pod "pod1-1-sched-preemption-medium-priority" in namespace "sched-preemption-8428" to be "running" -Dec 14 16:26:13.709: INFO: Pod "pod1-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 7.485337ms -Dec 14 16:26:13.709: INFO: Pod "pod1-1-sched-preemption-medium-priority" satisfied condition "running" -Dec 14 16:26:13.709: INFO: Waiting up to 5m0s for pod "pod2-0-sched-preemption-medium-priority" in namespace "sched-preemption-8428" to be "running" -Dec 14 16:26:13.714: INFO: Pod "pod2-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 4.724829ms -Dec 14 16:26:13.714: INFO: Pod "pod2-0-sched-preemption-medium-priority" satisfied condition "running" -Dec 14 16:26:13.714: INFO: Waiting up to 5m0s for pod "pod2-1-sched-preemption-medium-priority" in namespace "sched-preemption-8428" to be "running" -Dec 14 16:26:13.720: INFO: Pod "pod2-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 5.217729ms -Dec 14 16:26:13.720: INFO: Pod "pod2-1-sched-preemption-medium-priority" satisfied condition "running" -STEP: Run a critical pod that use same resources as that of a lower priority pod 12/14/22 16:26:13.72 -Dec 14 16:26:13.740: INFO: Waiting up to 2m0s for pod "critical-pod" in namespace "kube-system" to be "running" -Dec 14 16:26:13.747: INFO: Pod "critical-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 6.523995ms -Dec 14 16:26:15.763: INFO: Pod "critical-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022419828s -Dec 14 16:26:17.760: INFO: Pod "critical-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.019343828s -Dec 14 16:26:17.760: INFO: Pod "critical-pod" satisfied condition "running" -[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] +STEP: Creating a kubernetes client 02/25/23 10:23:21.829 +Feb 25 10:23:21.829: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename security-context 02/25/23 10:23:21.832 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:23:21.864 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:23:21.868 +[It] should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] + test/e2e/node/security_context.go:132 +STEP: Creating a pod to test pod.Spec.SecurityContext.RunAsUser 02/25/23 10:23:21.873 +Feb 25 10:23:21.888: INFO: Waiting up to 5m0s for pod "security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd" in namespace "security-context-826" to be "Succeeded or Failed" +Feb 25 10:23:21.896: INFO: Pod "security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd": Phase="Pending", Reason="", readiness=false. Elapsed: 8.767027ms +Feb 25 10:23:23.913: INFO: Pod "security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025784701s +Feb 25 10:23:25.905: INFO: Pod "security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd": Phase="Pending", Reason="", readiness=false. Elapsed: 4.017457267s +Feb 25 10:23:27.910: INFO: Pod "security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.022313451s +STEP: Saw pod success 02/25/23 10:23:27.91 +Feb 25 10:23:27.910: INFO: Pod "security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd" satisfied condition "Succeeded or Failed" +Feb 25 10:23:27.920: INFO: Trying to get logs from node diek3dua9gay-3 pod security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd container test-container: +STEP: delete the pod 02/25/23 10:23:27.978 +Feb 25 10:23:28.012: INFO: Waiting for pod security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd to disappear +Feb 25 10:23:28.023: INFO: Pod security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd no longer exists +[AfterEach] [sig-node] Security Context test/e2e/framework/framework.go:187 -Dec 14 16:26:17.836: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "sched-preemption-8428" for this suite. 12/14/22 16:26:17.841 -[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:80 -{"msg":"PASSED [sig-scheduling] SchedulerPreemption [Serial] validates lower priority pod preemption by critical pod [Conformance]","completed":136,"skipped":2445,"failed":0} +Feb 25 10:23:28.024: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "security-context-826" for this suite. 02/25/23 10:23:28.037 +{"msg":"PASSED [sig-node] Security Context should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance]","completed":130,"skipped":2431,"failed":0} ------------------------------ -• [SLOW TEST] [76.532 seconds] -[sig-scheduling] SchedulerPreemption [Serial] -test/e2e/scheduling/framework.go:40 - validates lower priority pod preemption by critical pod [Conformance] - test/e2e/scheduling/preemption.go:218 +• [SLOW TEST] [6.230 seconds] +[sig-node] Security Context +test/e2e/node/framework.go:23 + should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] + test/e2e/node/security_context.go:132 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] + [BeforeEach] [sig-node] Security Context test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:25:01.402 - Dec 14 16:25:01.402: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename sched-preemption 12/14/22 16:25:01.406 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:25:01.46 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:25:01.466 - [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:92 - Dec 14 16:25:01.494: INFO: Waiting up to 1m0s for all nodes to be ready - Dec 14 16:26:01.535: INFO: Waiting for terminating namespaces to be deleted... - [It] validates lower priority pod preemption by critical pod [Conformance] - test/e2e/scheduling/preemption.go:218 - STEP: Create pods that use 4/5 of node resources. 12/14/22 16:26:01.541 - Dec 14 16:26:01.580: INFO: Created pod: pod0-0-sched-preemption-low-priority - Dec 14 16:26:01.589: INFO: Created pod: pod0-1-sched-preemption-medium-priority - Dec 14 16:26:01.614: INFO: Created pod: pod1-0-sched-preemption-medium-priority - Dec 14 16:26:01.621: INFO: Created pod: pod1-1-sched-preemption-medium-priority - Dec 14 16:26:01.648: INFO: Created pod: pod2-0-sched-preemption-medium-priority - Dec 14 16:26:01.658: INFO: Created pod: pod2-1-sched-preemption-medium-priority - STEP: Wait for pods to be scheduled. 12/14/22 16:26:01.658 - Dec 14 16:26:01.659: INFO: Waiting up to 5m0s for pod "pod0-0-sched-preemption-low-priority" in namespace "sched-preemption-8428" to be "running" - Dec 14 16:26:01.664: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 5.237849ms - Dec 14 16:26:03.674: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014987854s - Dec 14 16:26:05.674: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 4.01528966s - Dec 14 16:26:07.676: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Running", Reason="", readiness=true. Elapsed: 6.017220372s - Dec 14 16:26:07.676: INFO: Pod "pod0-0-sched-preemption-low-priority" satisfied condition "running" - Dec 14 16:26:07.676: INFO: Waiting up to 5m0s for pod "pod0-1-sched-preemption-medium-priority" in namespace "sched-preemption-8428" to be "running" - Dec 14 16:26:07.683: INFO: Pod "pod0-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 6.628224ms - Dec 14 16:26:07.683: INFO: Pod "pod0-1-sched-preemption-medium-priority" satisfied condition "running" - Dec 14 16:26:07.683: INFO: Waiting up to 5m0s for pod "pod1-0-sched-preemption-medium-priority" in namespace "sched-preemption-8428" to be "running" - Dec 14 16:26:07.690: INFO: Pod "pod1-0-sched-preemption-medium-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 6.501223ms - Dec 14 16:26:09.700: INFO: Pod "pod1-0-sched-preemption-medium-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017010901s - Dec 14 16:26:11.699: INFO: Pod "pod1-0-sched-preemption-medium-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 4.016155265s - Dec 14 16:26:13.701: INFO: Pod "pod1-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 6.017900562s - Dec 14 16:26:13.701: INFO: Pod "pod1-0-sched-preemption-medium-priority" satisfied condition "running" - Dec 14 16:26:13.701: INFO: Waiting up to 5m0s for pod "pod1-1-sched-preemption-medium-priority" in namespace "sched-preemption-8428" to be "running" - Dec 14 16:26:13.709: INFO: Pod "pod1-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 7.485337ms - Dec 14 16:26:13.709: INFO: Pod "pod1-1-sched-preemption-medium-priority" satisfied condition "running" - Dec 14 16:26:13.709: INFO: Waiting up to 5m0s for pod "pod2-0-sched-preemption-medium-priority" in namespace "sched-preemption-8428" to be "running" - Dec 14 16:26:13.714: INFO: Pod "pod2-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 4.724829ms - Dec 14 16:26:13.714: INFO: Pod "pod2-0-sched-preemption-medium-priority" satisfied condition "running" - Dec 14 16:26:13.714: INFO: Waiting up to 5m0s for pod "pod2-1-sched-preemption-medium-priority" in namespace "sched-preemption-8428" to be "running" - Dec 14 16:26:13.720: INFO: Pod "pod2-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 5.217729ms - Dec 14 16:26:13.720: INFO: Pod "pod2-1-sched-preemption-medium-priority" satisfied condition "running" - STEP: Run a critical pod that use same resources as that of a lower priority pod 12/14/22 16:26:13.72 - Dec 14 16:26:13.740: INFO: Waiting up to 2m0s for pod "critical-pod" in namespace "kube-system" to be "running" - Dec 14 16:26:13.747: INFO: Pod "critical-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 6.523995ms - Dec 14 16:26:15.763: INFO: Pod "critical-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022419828s - Dec 14 16:26:17.760: INFO: Pod "critical-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.019343828s - Dec 14 16:26:17.760: INFO: Pod "critical-pod" satisfied condition "running" - [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] + STEP: Creating a kubernetes client 02/25/23 10:23:21.829 + Feb 25 10:23:21.829: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename security-context 02/25/23 10:23:21.832 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:23:21.864 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:23:21.868 + [It] should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] + test/e2e/node/security_context.go:132 + STEP: Creating a pod to test pod.Spec.SecurityContext.RunAsUser 02/25/23 10:23:21.873 + Feb 25 10:23:21.888: INFO: Waiting up to 5m0s for pod "security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd" in namespace "security-context-826" to be "Succeeded or Failed" + Feb 25 10:23:21.896: INFO: Pod "security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd": Phase="Pending", Reason="", readiness=false. Elapsed: 8.767027ms + Feb 25 10:23:23.913: INFO: Pod "security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025784701s + Feb 25 10:23:25.905: INFO: Pod "security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd": Phase="Pending", Reason="", readiness=false. Elapsed: 4.017457267s + Feb 25 10:23:27.910: INFO: Pod "security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.022313451s + STEP: Saw pod success 02/25/23 10:23:27.91 + Feb 25 10:23:27.910: INFO: Pod "security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd" satisfied condition "Succeeded or Failed" + Feb 25 10:23:27.920: INFO: Trying to get logs from node diek3dua9gay-3 pod security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd container test-container: + STEP: delete the pod 02/25/23 10:23:27.978 + Feb 25 10:23:28.012: INFO: Waiting for pod security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd to disappear + Feb 25 10:23:28.023: INFO: Pod security-context-f49f841d-5921-4cdb-8f53-0d58d3f6aedd no longer exists + [AfterEach] [sig-node] Security Context test/e2e/framework/framework.go:187 - Dec 14 16:26:17.836: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "sched-preemption-8428" for this suite. 12/14/22 16:26:17.841 - [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:80 + Feb 25 10:23:28.024: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "security-context-826" for this suite. 02/25/23 10:23:28.037 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Projected configMap - should be consumable from pods in volume as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:73 -[BeforeEach] [sig-storage] Projected configMap +[sig-network] Networking Granular Checks: Pods + should function for intra-pod communication: udp [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:93 +[BeforeEach] [sig-network] Networking test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:26:17.941 -Dec 14 16:26:17.941: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 16:26:17.944 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:26:17.967 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:26:17.974 -[It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:73 -STEP: Creating configMap with name projected-configmap-test-volume-e29c57bf-88a7-4bc7-8eff-2f35a6bf0b16 12/14/22 16:26:17.981 -STEP: Creating a pod to test consume configMaps 12/14/22 16:26:17.996 -Dec 14 16:26:18.019: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637" in namespace "projected-4834" to be "Succeeded or Failed" -Dec 14 16:26:18.028: INFO: Pod "pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637": Phase="Pending", Reason="", readiness=false. Elapsed: 7.999541ms -Dec 14 16:26:20.034: INFO: Pod "pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01451341s -Dec 14 16:26:22.035: INFO: Pod "pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014870327s -STEP: Saw pod success 12/14/22 16:26:22.035 -Dec 14 16:26:22.035: INFO: Pod "pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637" satisfied condition "Succeeded or Failed" -Dec 14 16:26:22.040: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637 container agnhost-container: -STEP: delete the pod 12/14/22 16:26:22.097 -Dec 14 16:26:22.119: INFO: Waiting for pod pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637 to disappear -Dec 14 16:26:22.124: INFO: Pod pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637 no longer exists -[AfterEach] [sig-storage] Projected configMap +STEP: Creating a kubernetes client 02/25/23 10:23:28.072 +Feb 25 10:23:28.072: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename pod-network-test 02/25/23 10:23:28.074 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:23:28.107 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:23:28.114 +[It] should function for intra-pod communication: udp [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:93 +STEP: Performing setup for networking test in namespace pod-network-test-7294 02/25/23 10:23:28.121 +STEP: creating a selector 02/25/23 10:23:28.121 +STEP: Creating the service pods in kubernetes 02/25/23 10:23:28.121 +Feb 25 10:23:28.122: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable +Feb 25 10:23:28.260: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-7294" to be "running and ready" +Feb 25 10:23:28.315: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 55.770646ms +Feb 25 10:23:28.316: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:23:30.327: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.067205268s +Feb 25 10:23:30.327: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:23:32.324: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.064713147s +Feb 25 10:23:32.325: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 10:23:34.325: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.065371421s +Feb 25 10:23:34.325: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 10:23:36.328: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.068044943s +Feb 25 10:23:36.328: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 10:23:38.325: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.065319594s +Feb 25 10:23:38.325: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 10:23:40.325: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 12.065086633s +Feb 25 10:23:40.325: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 10:23:42.328: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 14.068035541s +Feb 25 10:23:42.328: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 10:23:44.322: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 16.062506043s +Feb 25 10:23:44.322: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 10:23:46.324: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 18.064348578s +Feb 25 10:23:46.324: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 10:23:48.325: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 20.065091169s +Feb 25 10:23:48.325: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 10:23:50.323: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 22.063473252s +Feb 25 10:23:50.323: INFO: The phase of Pod netserver-0 is Running (Ready = true) +Feb 25 10:23:50.323: INFO: Pod "netserver-0" satisfied condition "running and ready" +Feb 25 10:23:50.330: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-7294" to be "running and ready" +Feb 25 10:23:50.348: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 17.73489ms +Feb 25 10:23:50.348: INFO: The phase of Pod netserver-1 is Running (Ready = true) +Feb 25 10:23:50.348: INFO: Pod "netserver-1" satisfied condition "running and ready" +Feb 25 10:23:50.359: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-7294" to be "running and ready" +Feb 25 10:23:50.378: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 19.237771ms +Feb 25 10:23:50.379: INFO: The phase of Pod netserver-2 is Running (Ready = true) +Feb 25 10:23:50.379: INFO: Pod "netserver-2" satisfied condition "running and ready" +STEP: Creating test pods 02/25/23 10:23:50.385 +Feb 25 10:23:50.403: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-7294" to be "running" +Feb 25 10:23:50.409: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 5.685759ms +Feb 25 10:23:52.429: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.024976899s +Feb 25 10:23:52.430: INFO: Pod "test-container-pod" satisfied condition "running" +Feb 25 10:23:52.438: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 +Feb 25 10:23:52.438: INFO: Breadth first check of 10.233.65.117 on host 192.168.121.250... +Feb 25 10:23:52.450: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.99:9080/dial?request=hostname&protocol=udp&host=10.233.65.117&port=8081&tries=1'] Namespace:pod-network-test-7294 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:23:52.451: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:23:52.455: INFO: ExecWithOptions: Clientset creation +Feb 25 10:23:52.455: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-7294/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.99%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dudp%26host%3D10.233.65.117%26port%3D8081%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) +Feb 25 10:23:52.742: INFO: Waiting for responses: map[] +Feb 25 10:23:52.743: INFO: reached 10.233.65.117 after 0/1 tries +Feb 25 10:23:52.743: INFO: Breadth first check of 10.233.64.136 on host 192.168.121.117... +Feb 25 10:23:52.751: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.99:9080/dial?request=hostname&protocol=udp&host=10.233.64.136&port=8081&tries=1'] Namespace:pod-network-test-7294 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:23:52.751: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:23:52.752: INFO: ExecWithOptions: Clientset creation +Feb 25 10:23:52.753: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-7294/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.99%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dudp%26host%3D10.233.64.136%26port%3D8081%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) +Feb 25 10:23:52.898: INFO: Waiting for responses: map[] +Feb 25 10:23:52.898: INFO: reached 10.233.64.136 after 0/1 tries +Feb 25 10:23:52.898: INFO: Breadth first check of 10.233.66.113 on host 192.168.121.44... +Feb 25 10:23:52.918: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.99:9080/dial?request=hostname&protocol=udp&host=10.233.66.113&port=8081&tries=1'] Namespace:pod-network-test-7294 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 10:23:52.918: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 10:23:52.920: INFO: ExecWithOptions: Clientset creation +Feb 25 10:23:52.920: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-7294/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.99%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dudp%26host%3D10.233.66.113%26port%3D8081%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) +Feb 25 10:23:53.058: INFO: Waiting for responses: map[] +Feb 25 10:23:53.058: INFO: reached 10.233.66.113 after 0/1 tries +Feb 25 10:23:53.058: INFO: Going to retry 0 out of 3 pods.... +[AfterEach] [sig-network] Networking test/e2e/framework/framework.go:187 -Dec 14 16:26:22.124: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-4834" for this suite. 12/14/22 16:26:22.129 -{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]","completed":137,"skipped":2465,"failed":0} +Feb 25 10:23:53.059: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pod-network-test-7294" for this suite. 02/25/23 10:23:53.077 +{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [NodeConformance] [Conformance]","completed":131,"skipped":2457,"failed":0} ------------------------------ -• [4.199 seconds] -[sig-storage] Projected configMap -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:73 +• [SLOW TEST] [25.116 seconds] +[sig-network] Networking +test/e2e/common/network/framework.go:23 + Granular Checks: Pods + test/e2e/common/network/networking.go:32 + should function for intra-pod communication: udp [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:93 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected configMap + [BeforeEach] [sig-network] Networking test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:26:17.941 - Dec 14 16:26:17.941: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 16:26:17.944 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:26:17.967 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:26:17.974 - [It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:73 - STEP: Creating configMap with name projected-configmap-test-volume-e29c57bf-88a7-4bc7-8eff-2f35a6bf0b16 12/14/22 16:26:17.981 - STEP: Creating a pod to test consume configMaps 12/14/22 16:26:17.996 - Dec 14 16:26:18.019: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637" in namespace "projected-4834" to be "Succeeded or Failed" - Dec 14 16:26:18.028: INFO: Pod "pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637": Phase="Pending", Reason="", readiness=false. Elapsed: 7.999541ms - Dec 14 16:26:20.034: INFO: Pod "pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01451341s - Dec 14 16:26:22.035: INFO: Pod "pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014870327s - STEP: Saw pod success 12/14/22 16:26:22.035 - Dec 14 16:26:22.035: INFO: Pod "pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637" satisfied condition "Succeeded or Failed" - Dec 14 16:26:22.040: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637 container agnhost-container: - STEP: delete the pod 12/14/22 16:26:22.097 - Dec 14 16:26:22.119: INFO: Waiting for pod pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637 to disappear - Dec 14 16:26:22.124: INFO: Pod pod-projected-configmaps-6399941b-2e79-4b50-ab50-087a8ab89637 no longer exists - [AfterEach] [sig-storage] Projected configMap + STEP: Creating a kubernetes client 02/25/23 10:23:28.072 + Feb 25 10:23:28.072: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename pod-network-test 02/25/23 10:23:28.074 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:23:28.107 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:23:28.114 + [It] should function for intra-pod communication: udp [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:93 + STEP: Performing setup for networking test in namespace pod-network-test-7294 02/25/23 10:23:28.121 + STEP: creating a selector 02/25/23 10:23:28.121 + STEP: Creating the service pods in kubernetes 02/25/23 10:23:28.121 + Feb 25 10:23:28.122: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable + Feb 25 10:23:28.260: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-7294" to be "running and ready" + Feb 25 10:23:28.315: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 55.770646ms + Feb 25 10:23:28.316: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:23:30.327: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.067205268s + Feb 25 10:23:30.327: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:23:32.324: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.064713147s + Feb 25 10:23:32.325: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 10:23:34.325: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.065371421s + Feb 25 10:23:34.325: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 10:23:36.328: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.068044943s + Feb 25 10:23:36.328: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 10:23:38.325: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.065319594s + Feb 25 10:23:38.325: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 10:23:40.325: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 12.065086633s + Feb 25 10:23:40.325: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 10:23:42.328: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 14.068035541s + Feb 25 10:23:42.328: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 10:23:44.322: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 16.062506043s + Feb 25 10:23:44.322: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 10:23:46.324: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 18.064348578s + Feb 25 10:23:46.324: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 10:23:48.325: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 20.065091169s + Feb 25 10:23:48.325: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 10:23:50.323: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 22.063473252s + Feb 25 10:23:50.323: INFO: The phase of Pod netserver-0 is Running (Ready = true) + Feb 25 10:23:50.323: INFO: Pod "netserver-0" satisfied condition "running and ready" + Feb 25 10:23:50.330: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-7294" to be "running and ready" + Feb 25 10:23:50.348: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 17.73489ms + Feb 25 10:23:50.348: INFO: The phase of Pod netserver-1 is Running (Ready = true) + Feb 25 10:23:50.348: INFO: Pod "netserver-1" satisfied condition "running and ready" + Feb 25 10:23:50.359: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-7294" to be "running and ready" + Feb 25 10:23:50.378: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 19.237771ms + Feb 25 10:23:50.379: INFO: The phase of Pod netserver-2 is Running (Ready = true) + Feb 25 10:23:50.379: INFO: Pod "netserver-2" satisfied condition "running and ready" + STEP: Creating test pods 02/25/23 10:23:50.385 + Feb 25 10:23:50.403: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-7294" to be "running" + Feb 25 10:23:50.409: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 5.685759ms + Feb 25 10:23:52.429: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.024976899s + Feb 25 10:23:52.430: INFO: Pod "test-container-pod" satisfied condition "running" + Feb 25 10:23:52.438: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 + Feb 25 10:23:52.438: INFO: Breadth first check of 10.233.65.117 on host 192.168.121.250... + Feb 25 10:23:52.450: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.99:9080/dial?request=hostname&protocol=udp&host=10.233.65.117&port=8081&tries=1'] Namespace:pod-network-test-7294 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:23:52.451: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:23:52.455: INFO: ExecWithOptions: Clientset creation + Feb 25 10:23:52.455: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-7294/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.99%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dudp%26host%3D10.233.65.117%26port%3D8081%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) + Feb 25 10:23:52.742: INFO: Waiting for responses: map[] + Feb 25 10:23:52.743: INFO: reached 10.233.65.117 after 0/1 tries + Feb 25 10:23:52.743: INFO: Breadth first check of 10.233.64.136 on host 192.168.121.117... + Feb 25 10:23:52.751: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.99:9080/dial?request=hostname&protocol=udp&host=10.233.64.136&port=8081&tries=1'] Namespace:pod-network-test-7294 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:23:52.751: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:23:52.752: INFO: ExecWithOptions: Clientset creation + Feb 25 10:23:52.753: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-7294/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.99%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dudp%26host%3D10.233.64.136%26port%3D8081%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) + Feb 25 10:23:52.898: INFO: Waiting for responses: map[] + Feb 25 10:23:52.898: INFO: reached 10.233.64.136 after 0/1 tries + Feb 25 10:23:52.898: INFO: Breadth first check of 10.233.66.113 on host 192.168.121.44... + Feb 25 10:23:52.918: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.99:9080/dial?request=hostname&protocol=udp&host=10.233.66.113&port=8081&tries=1'] Namespace:pod-network-test-7294 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 10:23:52.918: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 10:23:52.920: INFO: ExecWithOptions: Clientset creation + Feb 25 10:23:52.920: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-7294/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.99%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dudp%26host%3D10.233.66.113%26port%3D8081%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) + Feb 25 10:23:53.058: INFO: Waiting for responses: map[] + Feb 25 10:23:53.058: INFO: reached 10.233.66.113 after 0/1 tries + Feb 25 10:23:53.058: INFO: Going to retry 0 out of 3 pods.... + [AfterEach] [sig-network] Networking test/e2e/framework/framework.go:187 - Dec 14 16:26:22.124: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-4834" for this suite. 12/14/22 16:26:22.129 + Feb 25 10:23:53.059: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "pod-network-test-7294" for this suite. 02/25/23 10:23:53.077 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSS +SSSSSSSSSSSS ------------------------------ [sig-storage] Downward API volume - should update annotations on modification [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:161 + should provide container's memory limit [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:206 [BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:26:22.143 -Dec 14 16:26:22.144: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 16:26:22.146 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:26:22.177 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:26:22.181 +STEP: Creating a kubernetes client 02/25/23 10:23:53.193 +Feb 25 10:23:53.194: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 10:23:53.196 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:23:53.303 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:23:53.31 [BeforeEach] [sig-storage] Downward API volume test/e2e/common/storage/downwardapi_volume.go:43 -[It] should update annotations on modification [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:161 -STEP: Creating the pod 12/14/22 16:26:22.185 -Dec 14 16:26:22.199: INFO: Waiting up to 5m0s for pod "annotationupdate01875680-143b-4c6c-a389-b2aec0636750" in namespace "downward-api-2510" to be "running and ready" -Dec 14 16:26:22.214: INFO: Pod "annotationupdate01875680-143b-4c6c-a389-b2aec0636750": Phase="Pending", Reason="", readiness=false. Elapsed: 14.115603ms -Dec 14 16:26:22.214: INFO: The phase of Pod annotationupdate01875680-143b-4c6c-a389-b2aec0636750 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:26:24.227: INFO: Pod "annotationupdate01875680-143b-4c6c-a389-b2aec0636750": Phase="Running", Reason="", readiness=true. Elapsed: 2.027296592s -Dec 14 16:26:24.227: INFO: The phase of Pod annotationupdate01875680-143b-4c6c-a389-b2aec0636750 is Running (Ready = true) -Dec 14 16:26:24.227: INFO: Pod "annotationupdate01875680-143b-4c6c-a389-b2aec0636750" satisfied condition "running and ready" -Dec 14 16:26:24.769: INFO: Successfully updated pod "annotationupdate01875680-143b-4c6c-a389-b2aec0636750" +[It] should provide container's memory limit [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:206 +STEP: Creating a pod to test downward API volume plugin 02/25/23 10:23:53.315 +Feb 25 10:23:53.421: INFO: Waiting up to 5m0s for pod "downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c" in namespace "downward-api-7527" to be "Succeeded or Failed" +Feb 25 10:23:53.430: INFO: Pod "downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c": Phase="Pending", Reason="", readiness=false. Elapsed: 8.300138ms +Feb 25 10:23:55.438: INFO: Pod "downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016158949s +Feb 25 10:23:57.440: INFO: Pod "downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018715969s +STEP: Saw pod success 02/25/23 10:23:57.44 +Feb 25 10:23:57.441: INFO: Pod "downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c" satisfied condition "Succeeded or Failed" +Feb 25 10:23:57.447: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c container client-container: +STEP: delete the pod 02/25/23 10:23:57.464 +Feb 25 10:23:57.493: INFO: Waiting for pod downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c to disappear +Feb 25 10:23:57.499: INFO: Pod downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c no longer exists [AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 -Dec 14 16:26:28.810: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-2510" for this suite. 12/14/22 16:26:28.822 -{"msg":"PASSED [sig-storage] Downward API volume should update annotations on modification [NodeConformance] [Conformance]","completed":138,"skipped":2473,"failed":0} +Feb 25 10:23:57.500: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-7527" for this suite. 02/25/23 10:23:57.507 +{"msg":"PASSED [sig-storage] Downward API volume should provide container's memory limit [NodeConformance] [Conformance]","completed":132,"skipped":2469,"failed":0} ------------------------------ -• [SLOW TEST] [6.697 seconds] +• [4.330 seconds] [sig-storage] Downward API volume test/e2e/common/storage/framework.go:23 - should update annotations on modification [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:161 + should provide container's memory limit [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:206 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:26:22.143 - Dec 14 16:26:22.144: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 16:26:22.146 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:26:22.177 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:26:22.181 + STEP: Creating a kubernetes client 02/25/23 10:23:53.193 + Feb 25 10:23:53.194: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 10:23:53.196 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:23:53.303 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:23:53.31 [BeforeEach] [sig-storage] Downward API volume test/e2e/common/storage/downwardapi_volume.go:43 - [It] should update annotations on modification [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:161 - STEP: Creating the pod 12/14/22 16:26:22.185 - Dec 14 16:26:22.199: INFO: Waiting up to 5m0s for pod "annotationupdate01875680-143b-4c6c-a389-b2aec0636750" in namespace "downward-api-2510" to be "running and ready" - Dec 14 16:26:22.214: INFO: Pod "annotationupdate01875680-143b-4c6c-a389-b2aec0636750": Phase="Pending", Reason="", readiness=false. Elapsed: 14.115603ms - Dec 14 16:26:22.214: INFO: The phase of Pod annotationupdate01875680-143b-4c6c-a389-b2aec0636750 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:26:24.227: INFO: Pod "annotationupdate01875680-143b-4c6c-a389-b2aec0636750": Phase="Running", Reason="", readiness=true. Elapsed: 2.027296592s - Dec 14 16:26:24.227: INFO: The phase of Pod annotationupdate01875680-143b-4c6c-a389-b2aec0636750 is Running (Ready = true) - Dec 14 16:26:24.227: INFO: Pod "annotationupdate01875680-143b-4c6c-a389-b2aec0636750" satisfied condition "running and ready" - Dec 14 16:26:24.769: INFO: Successfully updated pod "annotationupdate01875680-143b-4c6c-a389-b2aec0636750" + [It] should provide container's memory limit [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:206 + STEP: Creating a pod to test downward API volume plugin 02/25/23 10:23:53.315 + Feb 25 10:23:53.421: INFO: Waiting up to 5m0s for pod "downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c" in namespace "downward-api-7527" to be "Succeeded or Failed" + Feb 25 10:23:53.430: INFO: Pod "downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c": Phase="Pending", Reason="", readiness=false. Elapsed: 8.300138ms + Feb 25 10:23:55.438: INFO: Pod "downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016158949s + Feb 25 10:23:57.440: INFO: Pod "downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018715969s + STEP: Saw pod success 02/25/23 10:23:57.44 + Feb 25 10:23:57.441: INFO: Pod "downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c" satisfied condition "Succeeded or Failed" + Feb 25 10:23:57.447: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c container client-container: + STEP: delete the pod 02/25/23 10:23:57.464 + Feb 25 10:23:57.493: INFO: Waiting for pod downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c to disappear + Feb 25 10:23:57.499: INFO: Pod downwardapi-volume-42ddcd12-8c0f-4aeb-8894-559b7e9ee62c no longer exists [AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 - Dec 14 16:26:28.810: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-2510" for this suite. 12/14/22 16:26:28.822 - << End Captured GinkgoWriter Output ------------------------------- -SSS ------------------------------- -[sig-storage] Projected downwardAPI - should update annotations on modification [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:161 -[BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:26:28.842 -Dec 14 16:26:28.842: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 16:26:28.845 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:26:28.872 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:26:28.878 -[BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 -[It] should update annotations on modification [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:161 -STEP: Creating the pod 12/14/22 16:26:28.882 -Dec 14 16:26:28.896: INFO: Waiting up to 5m0s for pod "annotationupdatea967e588-e1cc-40d9-840e-a460b27889ef" in namespace "projected-2236" to be "running and ready" -Dec 14 16:26:28.900: INFO: Pod "annotationupdatea967e588-e1cc-40d9-840e-a460b27889ef": Phase="Pending", Reason="", readiness=false. Elapsed: 3.80896ms -Dec 14 16:26:28.900: INFO: The phase of Pod annotationupdatea967e588-e1cc-40d9-840e-a460b27889ef is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:26:30.910: INFO: Pod "annotationupdatea967e588-e1cc-40d9-840e-a460b27889ef": Phase="Running", Reason="", readiness=true. Elapsed: 2.013656983s -Dec 14 16:26:30.910: INFO: The phase of Pod annotationupdatea967e588-e1cc-40d9-840e-a460b27889ef is Running (Ready = true) -Dec 14 16:26:30.910: INFO: Pod "annotationupdatea967e588-e1cc-40d9-840e-a460b27889ef" satisfied condition "running and ready" -Dec 14 16:26:31.456: INFO: Successfully updated pod "annotationupdatea967e588-e1cc-40d9-840e-a460b27889ef" -[AfterEach] [sig-storage] Projected downwardAPI - test/e2e/framework/framework.go:187 -Dec 14 16:26:35.504: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-2236" for this suite. 12/14/22 16:26:35.513 -{"msg":"PASSED [sig-storage] Projected downwardAPI should update annotations on modification [NodeConformance] [Conformance]","completed":139,"skipped":2476,"failed":0} ------------------------------- -• [SLOW TEST] [6.681 seconds] -[sig-storage] Projected downwardAPI -test/e2e/common/storage/framework.go:23 - should update annotations on modification [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:161 - - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:26:28.842 - Dec 14 16:26:28.842: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 16:26:28.845 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:26:28.872 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:26:28.878 - [BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 - [It] should update annotations on modification [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:161 - STEP: Creating the pod 12/14/22 16:26:28.882 - Dec 14 16:26:28.896: INFO: Waiting up to 5m0s for pod "annotationupdatea967e588-e1cc-40d9-840e-a460b27889ef" in namespace "projected-2236" to be "running and ready" - Dec 14 16:26:28.900: INFO: Pod "annotationupdatea967e588-e1cc-40d9-840e-a460b27889ef": Phase="Pending", Reason="", readiness=false. Elapsed: 3.80896ms - Dec 14 16:26:28.900: INFO: The phase of Pod annotationupdatea967e588-e1cc-40d9-840e-a460b27889ef is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:26:30.910: INFO: Pod "annotationupdatea967e588-e1cc-40d9-840e-a460b27889ef": Phase="Running", Reason="", readiness=true. Elapsed: 2.013656983s - Dec 14 16:26:30.910: INFO: The phase of Pod annotationupdatea967e588-e1cc-40d9-840e-a460b27889ef is Running (Ready = true) - Dec 14 16:26:30.910: INFO: Pod "annotationupdatea967e588-e1cc-40d9-840e-a460b27889ef" satisfied condition "running and ready" - Dec 14 16:26:31.456: INFO: Successfully updated pod "annotationupdatea967e588-e1cc-40d9-840e-a460b27889ef" - [AfterEach] [sig-storage] Projected downwardAPI - test/e2e/framework/framework.go:187 - Dec 14 16:26:35.504: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-2236" for this suite. 12/14/22 16:26:35.513 - << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - removes definition from spec when one version gets changed to not be served [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:441 -[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:26:35.53 -Dec 14 16:26:35.530: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 16:26:35.532 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:26:35.554 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:26:35.558 -[It] removes definition from spec when one version gets changed to not be served [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:441 -STEP: set up a multi version CRD 12/14/22 16:26:35.562 -Dec 14 16:26:35.564: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: mark a version not serverd 12/14/22 16:26:45.068 -STEP: check the unserved version gets removed 12/14/22 16:26:45.115 -STEP: check the other version is not changed 12/14/22 16:26:49.511 -[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - test/e2e/framework/framework.go:187 -Dec 14 16:26:56.996: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-publish-openapi-8044" for this suite. 12/14/22 16:26:57.019 -{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] removes definition from spec when one version gets changed to not be served [Conformance]","completed":140,"skipped":2507,"failed":0} ------------------------------- -• [SLOW TEST] [21.499 seconds] -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - removes definition from spec when one version gets changed to not be served [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:441 - - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:26:35.53 - Dec 14 16:26:35.530: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 16:26:35.532 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:26:35.554 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:26:35.558 - [It] removes definition from spec when one version gets changed to not be served [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:441 - STEP: set up a multi version CRD 12/14/22 16:26:35.562 - Dec 14 16:26:35.564: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: mark a version not serverd 12/14/22 16:26:45.068 - STEP: check the unserved version gets removed 12/14/22 16:26:45.115 - STEP: check the other version is not changed 12/14/22 16:26:49.511 - [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - test/e2e/framework/framework.go:187 - Dec 14 16:26:56.996: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "crd-publish-openapi-8044" for this suite. 12/14/22 16:26:57.019 + Feb 25 10:23:57.500: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-7527" for this suite. 02/25/23 10:23:57.507 << End Captured GinkgoWriter Output ------------------------------ -SS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-network] DNS - should provide DNS for the cluster [Conformance] - test/e2e/network/dns.go:50 -[BeforeEach] [sig-network] DNS +[sig-api-machinery] server version + should find the server version [Conformance] + test/e2e/apimachinery/server_version.go:39 +[BeforeEach] [sig-api-machinery] server version test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:26:57.031 -Dec 14 16:26:57.032: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename dns 12/14/22 16:26:57.036 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:26:57.06 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:26:57.064 -[It] should provide DNS for the cluster [Conformance] - test/e2e/network/dns.go:50 -STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;sleep 1; done - 12/14/22 16:26:57.068 -STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;sleep 1; done - 12/14/22 16:26:57.069 -STEP: creating a pod to probe DNS 12/14/22 16:26:57.069 -STEP: submitting the pod to kubernetes 12/14/22 16:26:57.069 -Dec 14 16:26:57.083: INFO: Waiting up to 15m0s for pod "dns-test-94a07343-c1f0-4a25-87d4-570723a0fc25" in namespace "dns-5016" to be "running" -Dec 14 16:26:57.103: INFO: Pod "dns-test-94a07343-c1f0-4a25-87d4-570723a0fc25": Phase="Pending", Reason="", readiness=false. Elapsed: 20.499434ms -Dec 14 16:26:59.115: INFO: Pod "dns-test-94a07343-c1f0-4a25-87d4-570723a0fc25": Phase="Pending", Reason="", readiness=false. Elapsed: 2.032527336s -Dec 14 16:27:01.115: INFO: Pod "dns-test-94a07343-c1f0-4a25-87d4-570723a0fc25": Phase="Running", Reason="", readiness=true. Elapsed: 4.032352808s -Dec 14 16:27:01.115: INFO: Pod "dns-test-94a07343-c1f0-4a25-87d4-570723a0fc25" satisfied condition "running" -STEP: retrieving the pod 12/14/22 16:27:01.115 -STEP: looking for the results for each expected name from probers 12/14/22 16:27:01.121 -Dec 14 16:27:01.151: INFO: DNS probes using dns-5016/dns-test-94a07343-c1f0-4a25-87d4-570723a0fc25 succeeded - -STEP: deleting the pod 12/14/22 16:27:01.151 -[AfterEach] [sig-network] DNS +STEP: Creating a kubernetes client 02/25/23 10:23:57.529 +Feb 25 10:23:57.529: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename server-version 02/25/23 10:23:57.535 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:23:57.581 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:23:57.588 +[It] should find the server version [Conformance] + test/e2e/apimachinery/server_version.go:39 +STEP: Request ServerVersion 02/25/23 10:23:57.594 +STEP: Confirm major version 02/25/23 10:23:57.597 +Feb 25 10:23:57.597: INFO: Major version: 1 +STEP: Confirm minor version 02/25/23 10:23:57.597 +Feb 25 10:23:57.597: INFO: cleanMinorVersion: 25 +Feb 25 10:23:57.597: INFO: Minor version: 25 +[AfterEach] [sig-api-machinery] server version test/e2e/framework/framework.go:187 -Dec 14 16:27:01.179: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "dns-5016" for this suite. 12/14/22 16:27:01.187 -{"msg":"PASSED [sig-network] DNS should provide DNS for the cluster [Conformance]","completed":141,"skipped":2509,"failed":0} +Feb 25 10:23:57.597: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "server-version-910" for this suite. 02/25/23 10:23:57.624 +{"msg":"PASSED [sig-api-machinery] server version should find the server version [Conformance]","completed":133,"skipped":2504,"failed":0} ------------------------------ -• [4.163 seconds] -[sig-network] DNS -test/e2e/network/common/framework.go:23 - should provide DNS for the cluster [Conformance] - test/e2e/network/dns.go:50 +• [0.122 seconds] +[sig-api-machinery] server version +test/e2e/apimachinery/framework.go:23 + should find the server version [Conformance] + test/e2e/apimachinery/server_version.go:39 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] DNS + [BeforeEach] [sig-api-machinery] server version test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:26:57.031 - Dec 14 16:26:57.032: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename dns 12/14/22 16:26:57.036 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:26:57.06 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:26:57.064 - [It] should provide DNS for the cluster [Conformance] - test/e2e/network/dns.go:50 - STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;sleep 1; done - 12/14/22 16:26:57.068 - STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;sleep 1; done - 12/14/22 16:26:57.069 - STEP: creating a pod to probe DNS 12/14/22 16:26:57.069 - STEP: submitting the pod to kubernetes 12/14/22 16:26:57.069 - Dec 14 16:26:57.083: INFO: Waiting up to 15m0s for pod "dns-test-94a07343-c1f0-4a25-87d4-570723a0fc25" in namespace "dns-5016" to be "running" - Dec 14 16:26:57.103: INFO: Pod "dns-test-94a07343-c1f0-4a25-87d4-570723a0fc25": Phase="Pending", Reason="", readiness=false. Elapsed: 20.499434ms - Dec 14 16:26:59.115: INFO: Pod "dns-test-94a07343-c1f0-4a25-87d4-570723a0fc25": Phase="Pending", Reason="", readiness=false. Elapsed: 2.032527336s - Dec 14 16:27:01.115: INFO: Pod "dns-test-94a07343-c1f0-4a25-87d4-570723a0fc25": Phase="Running", Reason="", readiness=true. Elapsed: 4.032352808s - Dec 14 16:27:01.115: INFO: Pod "dns-test-94a07343-c1f0-4a25-87d4-570723a0fc25" satisfied condition "running" - STEP: retrieving the pod 12/14/22 16:27:01.115 - STEP: looking for the results for each expected name from probers 12/14/22 16:27:01.121 - Dec 14 16:27:01.151: INFO: DNS probes using dns-5016/dns-test-94a07343-c1f0-4a25-87d4-570723a0fc25 succeeded - - STEP: deleting the pod 12/14/22 16:27:01.151 - [AfterEach] [sig-network] DNS + STEP: Creating a kubernetes client 02/25/23 10:23:57.529 + Feb 25 10:23:57.529: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename server-version 02/25/23 10:23:57.535 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:23:57.581 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:23:57.588 + [It] should find the server version [Conformance] + test/e2e/apimachinery/server_version.go:39 + STEP: Request ServerVersion 02/25/23 10:23:57.594 + STEP: Confirm major version 02/25/23 10:23:57.597 + Feb 25 10:23:57.597: INFO: Major version: 1 + STEP: Confirm minor version 02/25/23 10:23:57.597 + Feb 25 10:23:57.597: INFO: cleanMinorVersion: 25 + Feb 25 10:23:57.597: INFO: Minor version: 25 + [AfterEach] [sig-api-machinery] server version test/e2e/framework/framework.go:187 - Dec 14 16:27:01.179: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "dns-5016" for this suite. 12/14/22 16:27:01.187 + Feb 25 10:23:57.597: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "server-version-910" for this suite. 02/25/23 10:23:57.624 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSS +SSSSSSS ------------------------------ -[sig-node] Security Context When creating a pod with privileged - should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:527 -[BeforeEach] [sig-node] Security Context +[sig-storage] Secrets + should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:46 +[BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:27:01.201 -Dec 14 16:27:01.202: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename security-context-test 12/14/22 16:27:01.204 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:01.231 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:01.234 -[BeforeEach] [sig-node] Security Context - test/e2e/common/node/security_context.go:49 -[It] should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:527 -Dec 14 16:27:01.253: INFO: Waiting up to 5m0s for pod "busybox-privileged-false-a0376f5a-129f-47f5-8835-87d5385620b6" in namespace "security-context-test-1253" to be "Succeeded or Failed" -Dec 14 16:27:01.259: INFO: Pod "busybox-privileged-false-a0376f5a-129f-47f5-8835-87d5385620b6": Phase="Pending", Reason="", readiness=false. Elapsed: 5.200324ms -Dec 14 16:27:03.267: INFO: Pod "busybox-privileged-false-a0376f5a-129f-47f5-8835-87d5385620b6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013840268s -Dec 14 16:27:05.268: INFO: Pod "busybox-privileged-false-a0376f5a-129f-47f5-8835-87d5385620b6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014477806s -Dec 14 16:27:05.268: INFO: Pod "busybox-privileged-false-a0376f5a-129f-47f5-8835-87d5385620b6" satisfied condition "Succeeded or Failed" -Dec 14 16:27:05.279: INFO: Got logs for pod "busybox-privileged-false-a0376f5a-129f-47f5-8835-87d5385620b6": "ip: RTNETLINK answers: Operation not permitted\n" -[AfterEach] [sig-node] Security Context +STEP: Creating a kubernetes client 02/25/23 10:23:57.652 +Feb 25 10:23:57.653: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename secrets 02/25/23 10:23:57.671 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:23:57.732 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:23:57.74 +[It] should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:46 +STEP: Creating secret with name secret-test-7eb80cfd-b0f7-4e62-b8cf-4709d6ea6038 02/25/23 10:23:57.748 +STEP: Creating a pod to test consume secrets 02/25/23 10:23:57.761 +Feb 25 10:23:57.780: INFO: Waiting up to 5m0s for pod "pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb" in namespace "secrets-414" to be "Succeeded or Failed" +Feb 25 10:23:57.794: INFO: Pod "pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb": Phase="Pending", Reason="", readiness=false. Elapsed: 13.763801ms +Feb 25 10:23:59.857: INFO: Pod "pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb": Phase="Pending", Reason="", readiness=false. Elapsed: 2.076194659s +Feb 25 10:24:01.802: INFO: Pod "pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021032714s +STEP: Saw pod success 02/25/23 10:24:01.802 +Feb 25 10:24:01.803: INFO: Pod "pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb" satisfied condition "Succeeded or Failed" +Feb 25 10:24:01.808: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb container secret-volume-test: +STEP: delete the pod 02/25/23 10:24:01.848 +Feb 25 10:24:01.867: INFO: Waiting for pod pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb to disappear +Feb 25 10:24:01.872: INFO: Pod pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb no longer exists +[AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 -Dec 14 16:27:05.279: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "security-context-test-1253" for this suite. 12/14/22 16:27:05.286 -{"msg":"PASSED [sig-node] Security Context When creating a pod with privileged should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]","completed":142,"skipped":2519,"failed":0} +Feb 25 10:24:01.873: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-414" for this suite. 02/25/23 10:24:01.881 +{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance]","completed":134,"skipped":2511,"failed":0} ------------------------------ -• [4.094 seconds] -[sig-node] Security Context -test/e2e/common/node/framework.go:23 - When creating a pod with privileged - test/e2e/common/node/security_context.go:490 - should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:527 +• [4.240 seconds] +[sig-storage] Secrets +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:46 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Security Context + [BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:27:01.201 - Dec 14 16:27:01.202: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename security-context-test 12/14/22 16:27:01.204 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:01.231 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:01.234 - [BeforeEach] [sig-node] Security Context - test/e2e/common/node/security_context.go:49 - [It] should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:527 - Dec 14 16:27:01.253: INFO: Waiting up to 5m0s for pod "busybox-privileged-false-a0376f5a-129f-47f5-8835-87d5385620b6" in namespace "security-context-test-1253" to be "Succeeded or Failed" - Dec 14 16:27:01.259: INFO: Pod "busybox-privileged-false-a0376f5a-129f-47f5-8835-87d5385620b6": Phase="Pending", Reason="", readiness=false. Elapsed: 5.200324ms - Dec 14 16:27:03.267: INFO: Pod "busybox-privileged-false-a0376f5a-129f-47f5-8835-87d5385620b6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013840268s - Dec 14 16:27:05.268: INFO: Pod "busybox-privileged-false-a0376f5a-129f-47f5-8835-87d5385620b6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014477806s - Dec 14 16:27:05.268: INFO: Pod "busybox-privileged-false-a0376f5a-129f-47f5-8835-87d5385620b6" satisfied condition "Succeeded or Failed" - Dec 14 16:27:05.279: INFO: Got logs for pod "busybox-privileged-false-a0376f5a-129f-47f5-8835-87d5385620b6": "ip: RTNETLINK answers: Operation not permitted\n" - [AfterEach] [sig-node] Security Context + STEP: Creating a kubernetes client 02/25/23 10:23:57.652 + Feb 25 10:23:57.653: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename secrets 02/25/23 10:23:57.671 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:23:57.732 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:23:57.74 + [It] should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:46 + STEP: Creating secret with name secret-test-7eb80cfd-b0f7-4e62-b8cf-4709d6ea6038 02/25/23 10:23:57.748 + STEP: Creating a pod to test consume secrets 02/25/23 10:23:57.761 + Feb 25 10:23:57.780: INFO: Waiting up to 5m0s for pod "pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb" in namespace "secrets-414" to be "Succeeded or Failed" + Feb 25 10:23:57.794: INFO: Pod "pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb": Phase="Pending", Reason="", readiness=false. Elapsed: 13.763801ms + Feb 25 10:23:59.857: INFO: Pod "pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb": Phase="Pending", Reason="", readiness=false. Elapsed: 2.076194659s + Feb 25 10:24:01.802: INFO: Pod "pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021032714s + STEP: Saw pod success 02/25/23 10:24:01.802 + Feb 25 10:24:01.803: INFO: Pod "pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb" satisfied condition "Succeeded or Failed" + Feb 25 10:24:01.808: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb container secret-volume-test: + STEP: delete the pod 02/25/23 10:24:01.848 + Feb 25 10:24:01.867: INFO: Waiting for pod pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb to disappear + Feb 25 10:24:01.872: INFO: Pod pod-secrets-d446d4bd-c7a9-4e5c-b234-7a23d9bf88bb no longer exists + [AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 - Dec 14 16:27:05.279: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "security-context-test-1253" for this suite. 12/14/22 16:27:05.286 + Feb 25 10:24:01.873: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "secrets-414" for this suite. 02/25/23 10:24:01.881 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSS ------------------------------ -[sig-storage] Subpath Atomic writer volumes - should support subpaths with projected pod [Conformance] - test/e2e/storage/subpath.go:106 -[BeforeEach] [sig-storage] Subpath +[sig-storage] Downward API volume + should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:83 +[BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:27:05.3 -Dec 14 16:27:05.300: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename subpath 12/14/22 16:27:05.302 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:05.331 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:05.335 -[BeforeEach] Atomic writer volumes - test/e2e/storage/subpath.go:40 -STEP: Setting up data 12/14/22 16:27:05.34 -[It] should support subpaths with projected pod [Conformance] - test/e2e/storage/subpath.go:106 -STEP: Creating pod pod-subpath-test-projected-tfn4 12/14/22 16:27:05.355 -STEP: Creating a pod to test atomic-volume-subpath 12/14/22 16:27:05.356 -Dec 14 16:27:05.377: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-tfn4" in namespace "subpath-9675" to be "Succeeded or Failed" -Dec 14 16:27:05.393: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Pending", Reason="", readiness=false. Elapsed: 15.916196ms -Dec 14 16:27:07.403: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 2.0258492s -Dec 14 16:27:09.402: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 4.024731136s -Dec 14 16:27:11.408: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 6.03096192s -Dec 14 16:27:13.401: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 8.024024083s -Dec 14 16:27:15.399: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 10.021603539s -Dec 14 16:27:17.402: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 12.024550394s -Dec 14 16:27:19.406: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 14.02925571s -Dec 14 16:27:21.402: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 16.025001229s -Dec 14 16:27:23.402: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 18.024411306s -Dec 14 16:27:25.403: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 20.02566551s -Dec 14 16:27:27.403: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 22.026162481s -Dec 14 16:27:29.405: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=false. Elapsed: 24.028303223s -Dec 14 16:27:31.400: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.022730351s -STEP: Saw pod success 12/14/22 16:27:31.4 -Dec 14 16:27:31.401: INFO: Pod "pod-subpath-test-projected-tfn4" satisfied condition "Succeeded or Failed" -Dec 14 16:27:31.407: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-subpath-test-projected-tfn4 container test-container-subpath-projected-tfn4: -STEP: delete the pod 12/14/22 16:27:31.421 -Dec 14 16:27:31.445: INFO: Waiting for pod pod-subpath-test-projected-tfn4 to disappear -Dec 14 16:27:31.450: INFO: Pod pod-subpath-test-projected-tfn4 no longer exists -STEP: Deleting pod pod-subpath-test-projected-tfn4 12/14/22 16:27:31.45 -Dec 14 16:27:31.451: INFO: Deleting pod "pod-subpath-test-projected-tfn4" in namespace "subpath-9675" -[AfterEach] [sig-storage] Subpath +STEP: Creating a kubernetes client 02/25/23 10:24:01.909 +Feb 25 10:24:01.910: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 10:24:01.914 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:24:01.935 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:24:01.939 +[BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 +[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:83 +STEP: Creating a pod to test downward API volume plugin 02/25/23 10:24:01.943 +Feb 25 10:24:01.964: INFO: Waiting up to 5m0s for pod "downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f" in namespace "downward-api-3414" to be "Succeeded or Failed" +Feb 25 10:24:01.972: INFO: Pod "downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f": Phase="Pending", Reason="", readiness=false. Elapsed: 7.603854ms +Feb 25 10:24:03.984: INFO: Pod "downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019054915s +Feb 25 10:24:05.980: INFO: Pod "downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015232521s +STEP: Saw pod success 02/25/23 10:24:05.98 +Feb 25 10:24:05.980: INFO: Pod "downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f" satisfied condition "Succeeded or Failed" +Feb 25 10:24:05.989: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f container client-container: +STEP: delete the pod 02/25/23 10:24:06.006 +Feb 25 10:24:06.032: INFO: Waiting for pod downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f to disappear +Feb 25 10:24:06.044: INFO: Pod downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f no longer exists +[AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 -Dec 14 16:27:31.456: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "subpath-9675" for this suite. 12/14/22 16:27:31.464 -{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with projected pod [Conformance]","completed":143,"skipped":2550,"failed":0} +Feb 25 10:24:06.044: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-3414" for this suite. 02/25/23 10:24:06.055 +{"msg":"PASSED [sig-storage] Downward API volume should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","completed":135,"skipped":2524,"failed":0} ------------------------------ -• [SLOW TEST] [26.172 seconds] -[sig-storage] Subpath -test/e2e/storage/utils/framework.go:23 - Atomic writer volumes - test/e2e/storage/subpath.go:36 - should support subpaths with projected pod [Conformance] - test/e2e/storage/subpath.go:106 +• [4.164 seconds] +[sig-storage] Downward API volume +test/e2e/common/storage/framework.go:23 + should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:83 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Subpath + [BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:27:05.3 - Dec 14 16:27:05.300: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename subpath 12/14/22 16:27:05.302 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:05.331 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:05.335 - [BeforeEach] Atomic writer volumes - test/e2e/storage/subpath.go:40 - STEP: Setting up data 12/14/22 16:27:05.34 - [It] should support subpaths with projected pod [Conformance] - test/e2e/storage/subpath.go:106 - STEP: Creating pod pod-subpath-test-projected-tfn4 12/14/22 16:27:05.355 - STEP: Creating a pod to test atomic-volume-subpath 12/14/22 16:27:05.356 - Dec 14 16:27:05.377: INFO: Waiting up to 5m0s for pod "pod-subpath-test-projected-tfn4" in namespace "subpath-9675" to be "Succeeded or Failed" - Dec 14 16:27:05.393: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Pending", Reason="", readiness=false. Elapsed: 15.916196ms - Dec 14 16:27:07.403: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 2.0258492s - Dec 14 16:27:09.402: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 4.024731136s - Dec 14 16:27:11.408: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 6.03096192s - Dec 14 16:27:13.401: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 8.024024083s - Dec 14 16:27:15.399: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 10.021603539s - Dec 14 16:27:17.402: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 12.024550394s - Dec 14 16:27:19.406: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 14.02925571s - Dec 14 16:27:21.402: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 16.025001229s - Dec 14 16:27:23.402: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 18.024411306s - Dec 14 16:27:25.403: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 20.02566551s - Dec 14 16:27:27.403: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=true. Elapsed: 22.026162481s - Dec 14 16:27:29.405: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Running", Reason="", readiness=false. Elapsed: 24.028303223s - Dec 14 16:27:31.400: INFO: Pod "pod-subpath-test-projected-tfn4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.022730351s - STEP: Saw pod success 12/14/22 16:27:31.4 - Dec 14 16:27:31.401: INFO: Pod "pod-subpath-test-projected-tfn4" satisfied condition "Succeeded or Failed" - Dec 14 16:27:31.407: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-subpath-test-projected-tfn4 container test-container-subpath-projected-tfn4: - STEP: delete the pod 12/14/22 16:27:31.421 - Dec 14 16:27:31.445: INFO: Waiting for pod pod-subpath-test-projected-tfn4 to disappear - Dec 14 16:27:31.450: INFO: Pod pod-subpath-test-projected-tfn4 no longer exists - STEP: Deleting pod pod-subpath-test-projected-tfn4 12/14/22 16:27:31.45 - Dec 14 16:27:31.451: INFO: Deleting pod "pod-subpath-test-projected-tfn4" in namespace "subpath-9675" - [AfterEach] [sig-storage] Subpath + STEP: Creating a kubernetes client 02/25/23 10:24:01.909 + Feb 25 10:24:01.910: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 10:24:01.914 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:24:01.935 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:24:01.939 + [BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 + [It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:83 + STEP: Creating a pod to test downward API volume plugin 02/25/23 10:24:01.943 + Feb 25 10:24:01.964: INFO: Waiting up to 5m0s for pod "downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f" in namespace "downward-api-3414" to be "Succeeded or Failed" + Feb 25 10:24:01.972: INFO: Pod "downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f": Phase="Pending", Reason="", readiness=false. Elapsed: 7.603854ms + Feb 25 10:24:03.984: INFO: Pod "downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019054915s + Feb 25 10:24:05.980: INFO: Pod "downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015232521s + STEP: Saw pod success 02/25/23 10:24:05.98 + Feb 25 10:24:05.980: INFO: Pod "downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f" satisfied condition "Succeeded or Failed" + Feb 25 10:24:05.989: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f container client-container: + STEP: delete the pod 02/25/23 10:24:06.006 + Feb 25 10:24:06.032: INFO: Waiting for pod downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f to disappear + Feb 25 10:24:06.044: INFO: Pod downwardapi-volume-2d1baca8-46d2-43f8-b2da-7a0118b5f56f no longer exists + [AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 - Dec 14 16:27:31.456: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "subpath-9675" for this suite. 12/14/22 16:27:31.464 + Feb 25 10:24:06.044: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-3414" for this suite. 02/25/23 10:24:06.055 << End Captured GinkgoWriter Output ------------------------------ -SS +SSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] Garbage collector - should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] - test/e2e/apimachinery/garbage_collector.go:550 -[BeforeEach] [sig-api-machinery] Garbage collector +[sig-apps] CronJob + should not schedule new jobs when ForbidConcurrent [Slow] [Conformance] + test/e2e/apps/cronjob.go:124 +[BeforeEach] [sig-apps] CronJob test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:27:31.473 -Dec 14 16:27:31.473: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename gc 12/14/22 16:27:31.475 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:31.501 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:31.506 -[It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] - test/e2e/apimachinery/garbage_collector.go:550 -STEP: create the deployment 12/14/22 16:27:31.51 -STEP: Wait for the Deployment to create new ReplicaSet 12/14/22 16:27:31.52 -STEP: delete the deployment 12/14/22 16:27:32.056 -STEP: wait for deployment deletion to see if the garbage collector mistakenly deletes the rs 12/14/22 16:27:32.076 -STEP: Gathering metrics 12/14/22 16:27:32.629 -Dec 14 16:27:32.670: INFO: Waiting up to 5m0s for pod "kube-controller-manager-kiedai4iakae-2" in namespace "kube-system" to be "running and ready" -Dec 14 16:27:32.676: INFO: Pod "kube-controller-manager-kiedai4iakae-2": Phase="Running", Reason="", readiness=true. Elapsed: 5.689668ms -Dec 14 16:27:32.676: INFO: The phase of Pod kube-controller-manager-kiedai4iakae-2 is Running (Ready = true) -Dec 14 16:27:32.676: INFO: Pod "kube-controller-manager-kiedai4iakae-2" satisfied condition "running and ready" -Dec 14 16:27:32.776: INFO: For apiserver_request_total: -For apiserver_request_latency_seconds: -For apiserver_init_events_total: -For garbage_collector_attempt_to_delete_queue_latency: -For garbage_collector_attempt_to_delete_work_duration: -For garbage_collector_attempt_to_orphan_queue_latency: -For garbage_collector_attempt_to_orphan_work_duration: -For garbage_collector_dirty_processing_latency_microseconds: -For garbage_collector_event_processing_latency_microseconds: -For garbage_collector_graph_changes_queue_latency: -For garbage_collector_graph_changes_work_duration: -For garbage_collector_orphan_processing_latency_microseconds: -For namespace_queue_latency: -For namespace_queue_latency_sum: -For namespace_queue_latency_count: -For namespace_retries: -For namespace_work_duration: -For namespace_work_duration_sum: -For namespace_work_duration_count: -For function_duration_seconds: -For errors_total: -For evicted_pods_total: - -[AfterEach] [sig-api-machinery] Garbage collector +STEP: Creating a kubernetes client 02/25/23 10:24:06.076 +Feb 25 10:24:06.076: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename cronjob 02/25/23 10:24:06.084 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:24:06.115 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:24:06.119 +[It] should not schedule new jobs when ForbidConcurrent [Slow] [Conformance] + test/e2e/apps/cronjob.go:124 +STEP: Creating a ForbidConcurrent cronjob 02/25/23 10:24:06.124 +STEP: Ensuring a job is scheduled 02/25/23 10:24:06.138 +STEP: Ensuring exactly one is scheduled 02/25/23 10:25:02.149 +STEP: Ensuring exactly one running job exists by listing jobs explicitly 02/25/23 10:25:02.161 +STEP: Ensuring no more jobs are scheduled 02/25/23 10:25:02.174 +STEP: Removing cronjob 02/25/23 10:30:02.198 +[AfterEach] [sig-apps] CronJob test/e2e/framework/framework.go:187 -Dec 14 16:27:32.776: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "gc-8808" for this suite. 12/14/22 16:27:32.788 -{"msg":"PASSED [sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]","completed":144,"skipped":2552,"failed":0} +Feb 25 10:30:02.257: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "cronjob-39" for this suite. 02/25/23 10:30:02.269 +{"msg":"PASSED [sig-apps] CronJob should not schedule new jobs when ForbidConcurrent [Slow] [Conformance]","completed":136,"skipped":2546,"failed":0} ------------------------------ -• [1.327 seconds] -[sig-api-machinery] Garbage collector -test/e2e/apimachinery/framework.go:23 - should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] - test/e2e/apimachinery/garbage_collector.go:550 +• [SLOW TEST] [356.230 seconds] +[sig-apps] CronJob +test/e2e/apps/framework.go:23 + should not schedule new jobs when ForbidConcurrent [Slow] [Conformance] + test/e2e/apps/cronjob.go:124 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Garbage collector + [BeforeEach] [sig-apps] CronJob test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:27:31.473 - Dec 14 16:27:31.473: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename gc 12/14/22 16:27:31.475 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:31.501 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:31.506 - [It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] - test/e2e/apimachinery/garbage_collector.go:550 - STEP: create the deployment 12/14/22 16:27:31.51 - STEP: Wait for the Deployment to create new ReplicaSet 12/14/22 16:27:31.52 - STEP: delete the deployment 12/14/22 16:27:32.056 - STEP: wait for deployment deletion to see if the garbage collector mistakenly deletes the rs 12/14/22 16:27:32.076 - STEP: Gathering metrics 12/14/22 16:27:32.629 - Dec 14 16:27:32.670: INFO: Waiting up to 5m0s for pod "kube-controller-manager-kiedai4iakae-2" in namespace "kube-system" to be "running and ready" - Dec 14 16:27:32.676: INFO: Pod "kube-controller-manager-kiedai4iakae-2": Phase="Running", Reason="", readiness=true. Elapsed: 5.689668ms - Dec 14 16:27:32.676: INFO: The phase of Pod kube-controller-manager-kiedai4iakae-2 is Running (Ready = true) - Dec 14 16:27:32.676: INFO: Pod "kube-controller-manager-kiedai4iakae-2" satisfied condition "running and ready" - Dec 14 16:27:32.776: INFO: For apiserver_request_total: - For apiserver_request_latency_seconds: - For apiserver_init_events_total: - For garbage_collector_attempt_to_delete_queue_latency: - For garbage_collector_attempt_to_delete_work_duration: - For garbage_collector_attempt_to_orphan_queue_latency: - For garbage_collector_attempt_to_orphan_work_duration: - For garbage_collector_dirty_processing_latency_microseconds: - For garbage_collector_event_processing_latency_microseconds: - For garbage_collector_graph_changes_queue_latency: - For garbage_collector_graph_changes_work_duration: - For garbage_collector_orphan_processing_latency_microseconds: - For namespace_queue_latency: - For namespace_queue_latency_sum: - For namespace_queue_latency_count: - For namespace_retries: - For namespace_work_duration: - For namespace_work_duration_sum: - For namespace_work_duration_count: - For function_duration_seconds: - For errors_total: - For evicted_pods_total: - - [AfterEach] [sig-api-machinery] Garbage collector + STEP: Creating a kubernetes client 02/25/23 10:24:06.076 + Feb 25 10:24:06.076: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename cronjob 02/25/23 10:24:06.084 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:24:06.115 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:24:06.119 + [It] should not schedule new jobs when ForbidConcurrent [Slow] [Conformance] + test/e2e/apps/cronjob.go:124 + STEP: Creating a ForbidConcurrent cronjob 02/25/23 10:24:06.124 + STEP: Ensuring a job is scheduled 02/25/23 10:24:06.138 + STEP: Ensuring exactly one is scheduled 02/25/23 10:25:02.149 + STEP: Ensuring exactly one running job exists by listing jobs explicitly 02/25/23 10:25:02.161 + STEP: Ensuring no more jobs are scheduled 02/25/23 10:25:02.174 + STEP: Removing cronjob 02/25/23 10:30:02.198 + [AfterEach] [sig-apps] CronJob test/e2e/framework/framework.go:187 - Dec 14 16:27:32.776: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "gc-8808" for this suite. 12/14/22 16:27:32.788 + Feb 25 10:30:02.257: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "cronjob-39" for this suite. 02/25/23 10:30:02.269 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-instrumentation] Events API - should delete a collection of events [Conformance] - test/e2e/instrumentation/events.go:207 -[BeforeEach] [sig-instrumentation] Events API +[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook + should execute prestop http hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:152 +[BeforeEach] [sig-node] Container Lifecycle Hook test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:27:32.81 -Dec 14 16:27:32.810: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename events 12/14/22 16:27:32.812 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:32.839 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:32.843 -[BeforeEach] [sig-instrumentation] Events API - test/e2e/instrumentation/events.go:84 -[It] should delete a collection of events [Conformance] - test/e2e/instrumentation/events.go:207 -STEP: Create set of events 12/14/22 16:27:32.847 -STEP: get a list of Events with a label in the current namespace 12/14/22 16:27:32.871 -STEP: delete a list of events 12/14/22 16:27:32.879 -Dec 14 16:27:32.879: INFO: requesting DeleteCollection of events -STEP: check that the list of events matches the requested quantity 12/14/22 16:27:32.912 -[AfterEach] [sig-instrumentation] Events API +STEP: Creating a kubernetes client 02/25/23 10:30:02.316 +Feb 25 10:30:02.316: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-lifecycle-hook 02/25/23 10:30:02.321 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:02.397 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:02.407 +[BeforeEach] when create a pod with lifecycle hook + test/e2e/common/node/lifecycle_hook.go:55 +STEP: create the container to handle the HTTPGet hook request. 02/25/23 10:30:02.423 +Feb 25 10:30:02.441: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-4996" to be "running and ready" +Feb 25 10:30:02.455: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 13.570753ms +Feb 25 10:30:02.456: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:30:04.465: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023517408s +Feb 25 10:30:04.466: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:30:06.465: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 4.023597526s +Feb 25 10:30:06.465: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) +Feb 25 10:30:06.465: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" +[It] should execute prestop http hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:152 +STEP: create the pod with lifecycle hook 02/25/23 10:30:06.472 +Feb 25 10:30:06.484: INFO: Waiting up to 5m0s for pod "pod-with-prestop-http-hook" in namespace "container-lifecycle-hook-4996" to be "running and ready" +Feb 25 10:30:06.494: INFO: Pod "pod-with-prestop-http-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 10.467012ms +Feb 25 10:30:06.495: INFO: The phase of Pod pod-with-prestop-http-hook is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:30:08.503: INFO: Pod "pod-with-prestop-http-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018711701s +Feb 25 10:30:08.503: INFO: The phase of Pod pod-with-prestop-http-hook is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:30:10.503: INFO: Pod "pod-with-prestop-http-hook": Phase="Running", Reason="", readiness=true. Elapsed: 4.019250145s +Feb 25 10:30:10.503: INFO: The phase of Pod pod-with-prestop-http-hook is Running (Ready = true) +Feb 25 10:30:10.503: INFO: Pod "pod-with-prestop-http-hook" satisfied condition "running and ready" +STEP: delete the pod with lifecycle hook 02/25/23 10:30:10.511 +Feb 25 10:30:10.525: INFO: Waiting for pod pod-with-prestop-http-hook to disappear +Feb 25 10:30:10.560: INFO: Pod pod-with-prestop-http-hook still exists +Feb 25 10:30:12.560: INFO: Waiting for pod pod-with-prestop-http-hook to disappear +Feb 25 10:30:12.568: INFO: Pod pod-with-prestop-http-hook no longer exists +STEP: check prestop hook 02/25/23 10:30:12.568 +[AfterEach] [sig-node] Container Lifecycle Hook test/e2e/framework/framework.go:187 -Dec 14 16:27:32.916: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "events-4854" for this suite. 12/14/22 16:27:32.923 -{"msg":"PASSED [sig-instrumentation] Events API should delete a collection of events [Conformance]","completed":145,"skipped":2591,"failed":0} +Feb 25 10:30:12.608: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-lifecycle-hook-4996" for this suite. 02/25/23 10:30:12.62 +{"msg":"PASSED [sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance]","completed":137,"skipped":2574,"failed":0} ------------------------------ -• [0.124 seconds] -[sig-instrumentation] Events API -test/e2e/instrumentation/common/framework.go:23 - should delete a collection of events [Conformance] - test/e2e/instrumentation/events.go:207 +• [SLOW TEST] [10.322 seconds] +[sig-node] Container Lifecycle Hook +test/e2e/common/node/framework.go:23 + when create a pod with lifecycle hook + test/e2e/common/node/lifecycle_hook.go:46 + should execute prestop http hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:152 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-instrumentation] Events API + [BeforeEach] [sig-node] Container Lifecycle Hook test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:27:32.81 - Dec 14 16:27:32.810: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename events 12/14/22 16:27:32.812 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:32.839 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:32.843 - [BeforeEach] [sig-instrumentation] Events API - test/e2e/instrumentation/events.go:84 - [It] should delete a collection of events [Conformance] - test/e2e/instrumentation/events.go:207 - STEP: Create set of events 12/14/22 16:27:32.847 - STEP: get a list of Events with a label in the current namespace 12/14/22 16:27:32.871 - STEP: delete a list of events 12/14/22 16:27:32.879 - Dec 14 16:27:32.879: INFO: requesting DeleteCollection of events - STEP: check that the list of events matches the requested quantity 12/14/22 16:27:32.912 - [AfterEach] [sig-instrumentation] Events API + STEP: Creating a kubernetes client 02/25/23 10:30:02.316 + Feb 25 10:30:02.316: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-lifecycle-hook 02/25/23 10:30:02.321 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:02.397 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:02.407 + [BeforeEach] when create a pod with lifecycle hook + test/e2e/common/node/lifecycle_hook.go:55 + STEP: create the container to handle the HTTPGet hook request. 02/25/23 10:30:02.423 + Feb 25 10:30:02.441: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-4996" to be "running and ready" + Feb 25 10:30:02.455: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 13.570753ms + Feb 25 10:30:02.456: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:30:04.465: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023517408s + Feb 25 10:30:04.466: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:30:06.465: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 4.023597526s + Feb 25 10:30:06.465: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) + Feb 25 10:30:06.465: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" + [It] should execute prestop http hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:152 + STEP: create the pod with lifecycle hook 02/25/23 10:30:06.472 + Feb 25 10:30:06.484: INFO: Waiting up to 5m0s for pod "pod-with-prestop-http-hook" in namespace "container-lifecycle-hook-4996" to be "running and ready" + Feb 25 10:30:06.494: INFO: Pod "pod-with-prestop-http-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 10.467012ms + Feb 25 10:30:06.495: INFO: The phase of Pod pod-with-prestop-http-hook is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:30:08.503: INFO: Pod "pod-with-prestop-http-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018711701s + Feb 25 10:30:08.503: INFO: The phase of Pod pod-with-prestop-http-hook is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:30:10.503: INFO: Pod "pod-with-prestop-http-hook": Phase="Running", Reason="", readiness=true. Elapsed: 4.019250145s + Feb 25 10:30:10.503: INFO: The phase of Pod pod-with-prestop-http-hook is Running (Ready = true) + Feb 25 10:30:10.503: INFO: Pod "pod-with-prestop-http-hook" satisfied condition "running and ready" + STEP: delete the pod with lifecycle hook 02/25/23 10:30:10.511 + Feb 25 10:30:10.525: INFO: Waiting for pod pod-with-prestop-http-hook to disappear + Feb 25 10:30:10.560: INFO: Pod pod-with-prestop-http-hook still exists + Feb 25 10:30:12.560: INFO: Waiting for pod pod-with-prestop-http-hook to disappear + Feb 25 10:30:12.568: INFO: Pod pod-with-prestop-http-hook no longer exists + STEP: check prestop hook 02/25/23 10:30:12.568 + [AfterEach] [sig-node] Container Lifecycle Hook test/e2e/framework/framework.go:187 - Dec 14 16:27:32.916: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "events-4854" for this suite. 12/14/22 16:27:32.923 + Feb 25 10:30:12.608: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-lifecycle-hook-4996" for this suite. 02/25/23 10:30:12.62 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - should be able to deny pod and configmap creation [Conformance] - test/e2e/apimachinery/webhook.go:196 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +[sig-auth] ServiceAccounts + should allow opting out of API token automount [Conformance] + test/e2e/auth/service_accounts.go:158 +[BeforeEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:27:32.936 -Dec 14 16:27:32.936: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 16:27:32.938 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:32.959 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:32.962 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 16:27:32.99 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:27:33.502 -STEP: Deploying the webhook pod 12/14/22 16:27:33.547 -STEP: Wait for the deployment to be ready 12/14/22 16:27:33.568 -Dec 14 16:27:33.584: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set -STEP: Deploying the webhook service 12/14/22 16:27:35.606 -STEP: Verifying the service has paired with the endpoint 12/14/22 16:27:35.624 -Dec 14 16:27:36.624: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] should be able to deny pod and configmap creation [Conformance] - test/e2e/apimachinery/webhook.go:196 -STEP: Registering the webhook via the AdmissionRegistration API 12/14/22 16:27:36.63 -STEP: create a pod that should be denied by the webhook 12/14/22 16:27:36.674 -STEP: create a pod that causes the webhook to hang 12/14/22 16:27:36.701 -STEP: create a configmap that should be denied by the webhook 12/14/22 16:27:46.718 -STEP: create a configmap that should be admitted by the webhook 12/14/22 16:27:46.761 -STEP: update (PUT) the admitted configmap to a non-compliant one should be rejected by the webhook 12/14/22 16:27:46.791 -STEP: update (PATCH) the admitted configmap to a non-compliant one should be rejected by the webhook 12/14/22 16:27:46.812 -STEP: create a namespace that bypass the webhook 12/14/22 16:27:46.829 -STEP: create a configmap that violates the webhook policy but is in a whitelisted namespace 12/14/22 16:27:46.846 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 10:30:12.64 +Feb 25 10:30:12.641: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename svcaccounts 02/25/23 10:30:12.648 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:12.684 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:12.696 +[It] should allow opting out of API token automount [Conformance] + test/e2e/auth/service_accounts.go:158 +Feb 25 10:30:12.772: INFO: created pod pod-service-account-defaultsa +Feb 25 10:30:12.772: INFO: pod pod-service-account-defaultsa service account token volume mount: true +Feb 25 10:30:12.794: INFO: created pod pod-service-account-mountsa +Feb 25 10:30:12.794: INFO: pod pod-service-account-mountsa service account token volume mount: true +Feb 25 10:30:12.829: INFO: created pod pod-service-account-nomountsa +Feb 25 10:30:12.829: INFO: pod pod-service-account-nomountsa service account token volume mount: false +Feb 25 10:30:12.852: INFO: created pod pod-service-account-defaultsa-mountspec +Feb 25 10:30:12.853: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true +Feb 25 10:30:12.888: INFO: created pod pod-service-account-mountsa-mountspec +Feb 25 10:30:12.888: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true +Feb 25 10:30:12.959: INFO: created pod pod-service-account-nomountsa-mountspec +Feb 25 10:30:12.959: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true +Feb 25 10:30:13.036: INFO: created pod pod-service-account-defaultsa-nomountspec +Feb 25 10:30:13.036: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false +Feb 25 10:30:13.061: INFO: created pod pod-service-account-mountsa-nomountspec +Feb 25 10:30:13.061: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false +Feb 25 10:30:13.271: INFO: created pod pod-service-account-nomountsa-nomountspec +Feb 25 10:30:13.271: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false +[AfterEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:187 -Dec 14 16:27:46.882: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-2571" for this suite. 12/14/22 16:27:46.889 -STEP: Destroying namespace "webhook-2571-markers" for this suite. 12/14/22 16:27:46.898 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny pod and configmap creation [Conformance]","completed":146,"skipped":2610,"failed":0} +Feb 25 10:30:13.272: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "svcaccounts-3322" for this suite. 02/25/23 10:30:13.428 +{"msg":"PASSED [sig-auth] ServiceAccounts should allow opting out of API token automount [Conformance]","completed":138,"skipped":2574,"failed":0} ------------------------------ -• [SLOW TEST] [14.042 seconds] -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - should be able to deny pod and configmap creation [Conformance] - test/e2e/apimachinery/webhook.go:196 +• [0.949 seconds] +[sig-auth] ServiceAccounts +test/e2e/auth/framework.go:23 + should allow opting out of API token automount [Conformance] + test/e2e/auth/service_accounts.go:158 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + [BeforeEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:27:32.936 - Dec 14 16:27:32.936: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 16:27:32.938 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:32.959 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:32.962 - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 16:27:32.99 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:27:33.502 - STEP: Deploying the webhook pod 12/14/22 16:27:33.547 - STEP: Wait for the deployment to be ready 12/14/22 16:27:33.568 - Dec 14 16:27:33.584: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set - STEP: Deploying the webhook service 12/14/22 16:27:35.606 - STEP: Verifying the service has paired with the endpoint 12/14/22 16:27:35.624 - Dec 14 16:27:36.624: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] should be able to deny pod and configmap creation [Conformance] - test/e2e/apimachinery/webhook.go:196 - STEP: Registering the webhook via the AdmissionRegistration API 12/14/22 16:27:36.63 - STEP: create a pod that should be denied by the webhook 12/14/22 16:27:36.674 - STEP: create a pod that causes the webhook to hang 12/14/22 16:27:36.701 - STEP: create a configmap that should be denied by the webhook 12/14/22 16:27:46.718 - STEP: create a configmap that should be admitted by the webhook 12/14/22 16:27:46.761 - STEP: update (PUT) the admitted configmap to a non-compliant one should be rejected by the webhook 12/14/22 16:27:46.791 - STEP: update (PATCH) the admitted configmap to a non-compliant one should be rejected by the webhook 12/14/22 16:27:46.812 - STEP: create a namespace that bypass the webhook 12/14/22 16:27:46.829 - STEP: create a configmap that violates the webhook policy but is in a whitelisted namespace 12/14/22 16:27:46.846 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 10:30:12.64 + Feb 25 10:30:12.641: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename svcaccounts 02/25/23 10:30:12.648 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:12.684 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:12.696 + [It] should allow opting out of API token automount [Conformance] + test/e2e/auth/service_accounts.go:158 + Feb 25 10:30:12.772: INFO: created pod pod-service-account-defaultsa + Feb 25 10:30:12.772: INFO: pod pod-service-account-defaultsa service account token volume mount: true + Feb 25 10:30:12.794: INFO: created pod pod-service-account-mountsa + Feb 25 10:30:12.794: INFO: pod pod-service-account-mountsa service account token volume mount: true + Feb 25 10:30:12.829: INFO: created pod pod-service-account-nomountsa + Feb 25 10:30:12.829: INFO: pod pod-service-account-nomountsa service account token volume mount: false + Feb 25 10:30:12.852: INFO: created pod pod-service-account-defaultsa-mountspec + Feb 25 10:30:12.853: INFO: pod pod-service-account-defaultsa-mountspec service account token volume mount: true + Feb 25 10:30:12.888: INFO: created pod pod-service-account-mountsa-mountspec + Feb 25 10:30:12.888: INFO: pod pod-service-account-mountsa-mountspec service account token volume mount: true + Feb 25 10:30:12.959: INFO: created pod pod-service-account-nomountsa-mountspec + Feb 25 10:30:12.959: INFO: pod pod-service-account-nomountsa-mountspec service account token volume mount: true + Feb 25 10:30:13.036: INFO: created pod pod-service-account-defaultsa-nomountspec + Feb 25 10:30:13.036: INFO: pod pod-service-account-defaultsa-nomountspec service account token volume mount: false + Feb 25 10:30:13.061: INFO: created pod pod-service-account-mountsa-nomountspec + Feb 25 10:30:13.061: INFO: pod pod-service-account-mountsa-nomountspec service account token volume mount: false + Feb 25 10:30:13.271: INFO: created pod pod-service-account-nomountsa-nomountspec + Feb 25 10:30:13.271: INFO: pod pod-service-account-nomountsa-nomountspec service account token volume mount: false + [AfterEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:187 - Dec 14 16:27:46.882: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-2571" for this suite. 12/14/22 16:27:46.889 - STEP: Destroying namespace "webhook-2571-markers" for this suite. 12/14/22 16:27:46.898 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 + Feb 25 10:30:13.272: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "svcaccounts-3322" for this suite. 02/25/23 10:30:13.428 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSS ------------------------------ -[sig-node] PodTemplates - should run the lifecycle of PodTemplates [Conformance] - test/e2e/common/node/podtemplates.go:53 -[BeforeEach] [sig-node] PodTemplates +[sig-node] Variable Expansion + should fail substituting values in a volume subpath with backticks [Slow] [Conformance] + test/e2e/common/node/expansion.go:151 +[BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:27:46.988 -Dec 14 16:27:46.988: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename podtemplate 12/14/22 16:27:46.99 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:47.055 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:47.058 -[It] should run the lifecycle of PodTemplates [Conformance] - test/e2e/common/node/podtemplates.go:53 -[AfterEach] [sig-node] PodTemplates +STEP: Creating a kubernetes client 02/25/23 10:30:13.59 +Feb 25 10:30:13.643: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename var-expansion 02/25/23 10:30:13.672 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:13.736 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:13.751 +[It] should fail substituting values in a volume subpath with backticks [Slow] [Conformance] + test/e2e/common/node/expansion.go:151 +Feb 25 10:30:13.851: INFO: Waiting up to 2m0s for pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96" in namespace "var-expansion-6935" to be "container 0 failed with reason CreateContainerConfigError" +Feb 25 10:30:13.860: INFO: Pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96": Phase="Pending", Reason="", readiness=false. Elapsed: 9.020871ms +Feb 25 10:30:15.874: INFO: Pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022497422s +Feb 25 10:30:17.885: INFO: Pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96": Phase="Pending", Reason="", readiness=false. Elapsed: 4.033977654s +Feb 25 10:30:19.872: INFO: Pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96": Phase="Pending", Reason="", readiness=false. Elapsed: 6.020437256s +Feb 25 10:30:21.871: INFO: Pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96": Phase="Pending", Reason="", readiness=false. Elapsed: 8.01966921s +Feb 25 10:30:21.871: INFO: Pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96" satisfied condition "container 0 failed with reason CreateContainerConfigError" +Feb 25 10:30:21.871: INFO: Deleting pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96" in namespace "var-expansion-6935" +Feb 25 10:30:21.884: INFO: Wait up to 5m0s for pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96" to be fully deleted +[AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 -Dec 14 16:27:47.126: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "podtemplate-5834" for this suite. 12/14/22 16:27:47.135 -{"msg":"PASSED [sig-node] PodTemplates should run the lifecycle of PodTemplates [Conformance]","completed":147,"skipped":2670,"failed":0} +Feb 25 10:30:23.912: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "var-expansion-6935" for this suite. 02/25/23 10:30:23.921 +{"msg":"PASSED [sig-node] Variable Expansion should fail substituting values in a volume subpath with backticks [Slow] [Conformance]","completed":139,"skipped":2577,"failed":0} ------------------------------ -• [0.159 seconds] -[sig-node] PodTemplates +• [SLOW TEST] [10.345 seconds] +[sig-node] Variable Expansion test/e2e/common/node/framework.go:23 - should run the lifecycle of PodTemplates [Conformance] - test/e2e/common/node/podtemplates.go:53 + should fail substituting values in a volume subpath with backticks [Slow] [Conformance] + test/e2e/common/node/expansion.go:151 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] PodTemplates + [BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:27:46.988 - Dec 14 16:27:46.988: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename podtemplate 12/14/22 16:27:46.99 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:47.055 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:47.058 - [It] should run the lifecycle of PodTemplates [Conformance] - test/e2e/common/node/podtemplates.go:53 - [AfterEach] [sig-node] PodTemplates + STEP: Creating a kubernetes client 02/25/23 10:30:13.59 + Feb 25 10:30:13.643: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename var-expansion 02/25/23 10:30:13.672 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:13.736 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:13.751 + [It] should fail substituting values in a volume subpath with backticks [Slow] [Conformance] + test/e2e/common/node/expansion.go:151 + Feb 25 10:30:13.851: INFO: Waiting up to 2m0s for pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96" in namespace "var-expansion-6935" to be "container 0 failed with reason CreateContainerConfigError" + Feb 25 10:30:13.860: INFO: Pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96": Phase="Pending", Reason="", readiness=false. Elapsed: 9.020871ms + Feb 25 10:30:15.874: INFO: Pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022497422s + Feb 25 10:30:17.885: INFO: Pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96": Phase="Pending", Reason="", readiness=false. Elapsed: 4.033977654s + Feb 25 10:30:19.872: INFO: Pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96": Phase="Pending", Reason="", readiness=false. Elapsed: 6.020437256s + Feb 25 10:30:21.871: INFO: Pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96": Phase="Pending", Reason="", readiness=false. Elapsed: 8.01966921s + Feb 25 10:30:21.871: INFO: Pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96" satisfied condition "container 0 failed with reason CreateContainerConfigError" + Feb 25 10:30:21.871: INFO: Deleting pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96" in namespace "var-expansion-6935" + Feb 25 10:30:21.884: INFO: Wait up to 5m0s for pod "var-expansion-97abccb6-3d89-4cbb-a22c-8822855e9c96" to be fully deleted + [AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 - Dec 14 16:27:47.126: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "podtemplate-5834" for this suite. 12/14/22 16:27:47.135 + Feb 25 10:30:23.912: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "var-expansion-6935" for this suite. 02/25/23 10:30:23.921 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSS ------------------------------ -[sig-network] Services - should provide secure master service [Conformance] - test/e2e/network/service.go:781 -[BeforeEach] [sig-network] Services +[sig-storage] Downward API volume + should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:67 +[BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:27:47.154 -Dec 14 16:27:47.154: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 16:27:47.157 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:47.204 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:47.206 -[BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 -[It] should provide secure master service [Conformance] - test/e2e/network/service.go:781 -[AfterEach] [sig-network] Services +STEP: Creating a kubernetes client 02/25/23 10:30:23.937 +Feb 25 10:30:23.937: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 10:30:23.944 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:23.992 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:23.997 +[BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 +[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:67 +STEP: Creating a pod to test downward API volume plugin 02/25/23 10:30:24.003 +Feb 25 10:30:24.024: INFO: Waiting up to 5m0s for pod "downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef" in namespace "downward-api-1552" to be "Succeeded or Failed" +Feb 25 10:30:24.034: INFO: Pod "downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef": Phase="Pending", Reason="", readiness=false. Elapsed: 9.527844ms +Feb 25 10:30:26.044: INFO: Pod "downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01966933s +Feb 25 10:30:28.045: INFO: Pod "downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef": Phase="Running", Reason="", readiness=false. Elapsed: 4.020524071s +Feb 25 10:30:30.044: INFO: Pod "downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.019072088s +STEP: Saw pod success 02/25/23 10:30:30.044 +Feb 25 10:30:30.044: INFO: Pod "downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef" satisfied condition "Succeeded or Failed" +Feb 25 10:30:30.055: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef container client-container: +STEP: delete the pod 02/25/23 10:30:30.078 +Feb 25 10:30:30.113: INFO: Waiting for pod downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef to disappear +Feb 25 10:30:30.121: INFO: Pod downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef no longer exists +[AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 -Dec 14 16:27:47.214: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-4246" for this suite. 12/14/22 16:27:47.221 -[AfterEach] [sig-network] Services - test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should provide secure master service [Conformance]","completed":148,"skipped":2714,"failed":0} +Feb 25 10:30:30.122: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-1552" for this suite. 02/25/23 10:30:30.136 +{"msg":"PASSED [sig-storage] Downward API volume should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","completed":140,"skipped":2593,"failed":0} ------------------------------ -• [0.078 seconds] -[sig-network] Services -test/e2e/network/common/framework.go:23 - should provide secure master service [Conformance] - test/e2e/network/service.go:781 +• [SLOW TEST] [6.215 seconds] +[sig-storage] Downward API volume +test/e2e/common/storage/framework.go:23 + should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:67 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Services + [BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:27:47.154 - Dec 14 16:27:47.154: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 16:27:47.157 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:47.204 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:47.206 - [BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 - [It] should provide secure master service [Conformance] - test/e2e/network/service.go:781 - [AfterEach] [sig-network] Services + STEP: Creating a kubernetes client 02/25/23 10:30:23.937 + Feb 25 10:30:23.937: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 10:30:23.944 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:23.992 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:23.997 + [BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 + [It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:67 + STEP: Creating a pod to test downward API volume plugin 02/25/23 10:30:24.003 + Feb 25 10:30:24.024: INFO: Waiting up to 5m0s for pod "downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef" in namespace "downward-api-1552" to be "Succeeded or Failed" + Feb 25 10:30:24.034: INFO: Pod "downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef": Phase="Pending", Reason="", readiness=false. Elapsed: 9.527844ms + Feb 25 10:30:26.044: INFO: Pod "downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01966933s + Feb 25 10:30:28.045: INFO: Pod "downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef": Phase="Running", Reason="", readiness=false. Elapsed: 4.020524071s + Feb 25 10:30:30.044: INFO: Pod "downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.019072088s + STEP: Saw pod success 02/25/23 10:30:30.044 + Feb 25 10:30:30.044: INFO: Pod "downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef" satisfied condition "Succeeded or Failed" + Feb 25 10:30:30.055: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef container client-container: + STEP: delete the pod 02/25/23 10:30:30.078 + Feb 25 10:30:30.113: INFO: Waiting for pod downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef to disappear + Feb 25 10:30:30.121: INFO: Pod downwardapi-volume-10cfee75-99d0-4724-96cc-b548ebeae5ef no longer exists + [AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 - Dec 14 16:27:47.214: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-4246" for this suite. 12/14/22 16:27:47.221 - [AfterEach] [sig-network] Services - test/e2e/network/service.go:762 + Feb 25 10:30:30.122: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-1552" for this suite. 02/25/23 10:30:30.136 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSS ------------------------------ -[sig-node] Probing container - should have monotonically increasing restart count [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:195 -[BeforeEach] [sig-node] Probing container +[sig-node] Downward API + should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:165 +[BeforeEach] [sig-node] Downward API test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:27:47.238 -Dec 14 16:27:47.239: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-probe 12/14/22 16:27:47.242 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:47.289 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:47.293 -[BeforeEach] [sig-node] Probing container - test/e2e/common/node/container_probe.go:59 -[It] should have monotonically increasing restart count [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:195 -STEP: Creating pod liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 in namespace container-probe-457 12/14/22 16:27:47.297 -Dec 14 16:27:47.306: INFO: Waiting up to 5m0s for pod "liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5" in namespace "container-probe-457" to be "not pending" -Dec 14 16:27:47.329: INFO: Pod "liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5": Phase="Pending", Reason="", readiness=false. Elapsed: 22.992276ms -Dec 14 16:27:49.344: INFO: Pod "liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5": Phase="Running", Reason="", readiness=true. Elapsed: 2.037269832s -Dec 14 16:27:49.344: INFO: Pod "liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5" satisfied condition "not pending" -Dec 14 16:27:49.344: INFO: Started pod liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 in namespace container-probe-457 -STEP: checking the pod's current state and verifying that restartCount is present 12/14/22 16:27:49.344 -Dec 14 16:27:49.351: INFO: Initial restart count of pod liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 is 0 -Dec 14 16:28:09.443: INFO: Restart count of pod container-probe-457/liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 is now 1 (20.092718885s elapsed) -Dec 14 16:28:29.542: INFO: Restart count of pod container-probe-457/liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 is now 2 (40.191090005s elapsed) -Dec 14 16:28:49.655: INFO: Restart count of pod container-probe-457/liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 is now 3 (1m0.304409153s elapsed) -Dec 14 16:29:09.756: INFO: Restart count of pod container-probe-457/liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 is now 4 (1m20.404748412s elapsed) -Dec 14 16:30:16.056: INFO: Restart count of pod container-probe-457/liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 is now 5 (2m26.705606769s elapsed) -STEP: deleting the pod 12/14/22 16:30:16.057 -[AfterEach] [sig-node] Probing container +STEP: Creating a kubernetes client 02/25/23 10:30:30.171 +Feb 25 10:30:30.172: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 10:30:30.176 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:30.221 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:30.229 +[It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:165 +STEP: Creating a pod to test downward api env vars 02/25/23 10:30:30.236 +Feb 25 10:30:30.261: INFO: Waiting up to 5m0s for pod "downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24" in namespace "downward-api-674" to be "Succeeded or Failed" +Feb 25 10:30:30.268: INFO: Pod "downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24": Phase="Pending", Reason="", readiness=false. Elapsed: 6.546229ms +Feb 25 10:30:32.279: INFO: Pod "downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01790583s +Feb 25 10:30:34.274: INFO: Pod "downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012991529s +Feb 25 10:30:36.280: INFO: Pod "downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.018399264s +STEP: Saw pod success 02/25/23 10:30:36.28 +Feb 25 10:30:36.280: INFO: Pod "downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24" satisfied condition "Succeeded or Failed" +Feb 25 10:30:36.291: INFO: Trying to get logs from node diek3dua9gay-3 pod downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24 container dapi-container: +STEP: delete the pod 02/25/23 10:30:36.313 +Feb 25 10:30:36.352: INFO: Waiting for pod downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24 to disappear +Feb 25 10:30:36.363: INFO: Pod downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24 no longer exists +[AfterEach] [sig-node] Downward API test/e2e/framework/framework.go:187 -Dec 14 16:30:16.093: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-probe-457" for this suite. 12/14/22 16:30:16.102 -{"msg":"PASSED [sig-node] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance]","completed":149,"skipped":2743,"failed":0} +Feb 25 10:30:36.363: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-674" for this suite. 02/25/23 10:30:36.376 +{"msg":"PASSED [sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]","completed":141,"skipped":2602,"failed":0} ------------------------------ -• [SLOW TEST] [148.882 seconds] -[sig-node] Probing container +• [SLOW TEST] [6.224 seconds] +[sig-node] Downward API test/e2e/common/node/framework.go:23 - should have monotonically increasing restart count [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:195 + should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:165 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Probing container + [BeforeEach] [sig-node] Downward API test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:27:47.238 - Dec 14 16:27:47.239: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-probe 12/14/22 16:27:47.242 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:27:47.289 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:27:47.293 - [BeforeEach] [sig-node] Probing container - test/e2e/common/node/container_probe.go:59 - [It] should have monotonically increasing restart count [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:195 - STEP: Creating pod liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 in namespace container-probe-457 12/14/22 16:27:47.297 - Dec 14 16:27:47.306: INFO: Waiting up to 5m0s for pod "liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5" in namespace "container-probe-457" to be "not pending" - Dec 14 16:27:47.329: INFO: Pod "liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5": Phase="Pending", Reason="", readiness=false. Elapsed: 22.992276ms - Dec 14 16:27:49.344: INFO: Pod "liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5": Phase="Running", Reason="", readiness=true. Elapsed: 2.037269832s - Dec 14 16:27:49.344: INFO: Pod "liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5" satisfied condition "not pending" - Dec 14 16:27:49.344: INFO: Started pod liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 in namespace container-probe-457 - STEP: checking the pod's current state and verifying that restartCount is present 12/14/22 16:27:49.344 - Dec 14 16:27:49.351: INFO: Initial restart count of pod liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 is 0 - Dec 14 16:28:09.443: INFO: Restart count of pod container-probe-457/liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 is now 1 (20.092718885s elapsed) - Dec 14 16:28:29.542: INFO: Restart count of pod container-probe-457/liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 is now 2 (40.191090005s elapsed) - Dec 14 16:28:49.655: INFO: Restart count of pod container-probe-457/liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 is now 3 (1m0.304409153s elapsed) - Dec 14 16:29:09.756: INFO: Restart count of pod container-probe-457/liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 is now 4 (1m20.404748412s elapsed) - Dec 14 16:30:16.056: INFO: Restart count of pod container-probe-457/liveness-80d39ccb-a021-4c78-a777-95ec6c34aae5 is now 5 (2m26.705606769s elapsed) - STEP: deleting the pod 12/14/22 16:30:16.057 - [AfterEach] [sig-node] Probing container + STEP: Creating a kubernetes client 02/25/23 10:30:30.171 + Feb 25 10:30:30.172: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 10:30:30.176 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:30.221 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:30.229 + [It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:165 + STEP: Creating a pod to test downward api env vars 02/25/23 10:30:30.236 + Feb 25 10:30:30.261: INFO: Waiting up to 5m0s for pod "downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24" in namespace "downward-api-674" to be "Succeeded or Failed" + Feb 25 10:30:30.268: INFO: Pod "downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24": Phase="Pending", Reason="", readiness=false. Elapsed: 6.546229ms + Feb 25 10:30:32.279: INFO: Pod "downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01790583s + Feb 25 10:30:34.274: INFO: Pod "downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012991529s + Feb 25 10:30:36.280: INFO: Pod "downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.018399264s + STEP: Saw pod success 02/25/23 10:30:36.28 + Feb 25 10:30:36.280: INFO: Pod "downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24" satisfied condition "Succeeded or Failed" + Feb 25 10:30:36.291: INFO: Trying to get logs from node diek3dua9gay-3 pod downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24 container dapi-container: + STEP: delete the pod 02/25/23 10:30:36.313 + Feb 25 10:30:36.352: INFO: Waiting for pod downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24 to disappear + Feb 25 10:30:36.363: INFO: Pod downward-api-c245ee6c-4a0e-40b9-956b-17c237b29e24 no longer exists + [AfterEach] [sig-node] Downward API test/e2e/framework/framework.go:187 - Dec 14 16:30:16.093: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-probe-457" for this suite. 12/14/22 16:30:16.102 + Feb 25 10:30:36.363: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-674" for this suite. 02/25/23 10:30:36.376 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSS +SSSSS ------------------------------ -[sig-instrumentation] Events - should manage the lifecycle of an event [Conformance] - test/e2e/instrumentation/core_events.go:57 -[BeforeEach] [sig-instrumentation] Events +[sig-storage] Downward API volume + should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:260 +[BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:30:16.123 -Dec 14 16:30:16.123: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename events 12/14/22 16:30:16.128 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:16.173 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:16.178 -[It] should manage the lifecycle of an event [Conformance] - test/e2e/instrumentation/core_events.go:57 -STEP: creating a test event 12/14/22 16:30:16.182 -STEP: listing all events in all namespaces 12/14/22 16:30:16.198 -STEP: patching the test event 12/14/22 16:30:16.225 -STEP: fetching the test event 12/14/22 16:30:16.249 -STEP: updating the test event 12/14/22 16:30:16.256 -STEP: getting the test event 12/14/22 16:30:16.275 -STEP: deleting the test event 12/14/22 16:30:16.28 -STEP: listing all events in all namespaces 12/14/22 16:30:16.295 -[AfterEach] [sig-instrumentation] Events +STEP: Creating a kubernetes client 02/25/23 10:30:36.396 +Feb 25 10:30:36.396: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 10:30:36.4 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:36.434 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:36.44 +[BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 +[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:260 +STEP: Creating a pod to test downward API volume plugin 02/25/23 10:30:36.446 +Feb 25 10:30:36.464: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b" in namespace "downward-api-5158" to be "Succeeded or Failed" +Feb 25 10:30:36.481: INFO: Pod "downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b": Phase="Pending", Reason="", readiness=false. Elapsed: 16.871127ms +Feb 25 10:30:38.501: INFO: Pod "downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.036885299s +Feb 25 10:30:40.499: INFO: Pod "downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.035630957s +Feb 25 10:30:42.491: INFO: Pod "downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.027366818s +STEP: Saw pod success 02/25/23 10:30:42.491 +Feb 25 10:30:42.492: INFO: Pod "downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b" satisfied condition "Succeeded or Failed" +Feb 25 10:30:42.499: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b container client-container: +STEP: delete the pod 02/25/23 10:30:42.514 +Feb 25 10:30:42.557: INFO: Waiting for pod downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b to disappear +Feb 25 10:30:42.566: INFO: Pod downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b no longer exists +[AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 -Dec 14 16:30:16.324: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "events-7375" for this suite. 12/14/22 16:30:16.331 -{"msg":"PASSED [sig-instrumentation] Events should manage the lifecycle of an event [Conformance]","completed":150,"skipped":2752,"failed":0} +Feb 25 10:30:42.567: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-5158" for this suite. 02/25/23 10:30:42.648 +{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","completed":142,"skipped":2607,"failed":0} ------------------------------ -• [0.220 seconds] -[sig-instrumentation] Events -test/e2e/instrumentation/common/framework.go:23 - should manage the lifecycle of an event [Conformance] - test/e2e/instrumentation/core_events.go:57 +• [SLOW TEST] [6.265 seconds] +[sig-storage] Downward API volume +test/e2e/common/storage/framework.go:23 + should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:260 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-instrumentation] Events + [BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:30:16.123 - Dec 14 16:30:16.123: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename events 12/14/22 16:30:16.128 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:16.173 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:16.178 - [It] should manage the lifecycle of an event [Conformance] - test/e2e/instrumentation/core_events.go:57 - STEP: creating a test event 12/14/22 16:30:16.182 - STEP: listing all events in all namespaces 12/14/22 16:30:16.198 - STEP: patching the test event 12/14/22 16:30:16.225 - STEP: fetching the test event 12/14/22 16:30:16.249 - STEP: updating the test event 12/14/22 16:30:16.256 - STEP: getting the test event 12/14/22 16:30:16.275 - STEP: deleting the test event 12/14/22 16:30:16.28 - STEP: listing all events in all namespaces 12/14/22 16:30:16.295 - [AfterEach] [sig-instrumentation] Events + STEP: Creating a kubernetes client 02/25/23 10:30:36.396 + Feb 25 10:30:36.396: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 10:30:36.4 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:36.434 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:36.44 + [BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 + [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:260 + STEP: Creating a pod to test downward API volume plugin 02/25/23 10:30:36.446 + Feb 25 10:30:36.464: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b" in namespace "downward-api-5158" to be "Succeeded or Failed" + Feb 25 10:30:36.481: INFO: Pod "downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b": Phase="Pending", Reason="", readiness=false. Elapsed: 16.871127ms + Feb 25 10:30:38.501: INFO: Pod "downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.036885299s + Feb 25 10:30:40.499: INFO: Pod "downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.035630957s + Feb 25 10:30:42.491: INFO: Pod "downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.027366818s + STEP: Saw pod success 02/25/23 10:30:42.491 + Feb 25 10:30:42.492: INFO: Pod "downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b" satisfied condition "Succeeded or Failed" + Feb 25 10:30:42.499: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b container client-container: + STEP: delete the pod 02/25/23 10:30:42.514 + Feb 25 10:30:42.557: INFO: Waiting for pod downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b to disappear + Feb 25 10:30:42.566: INFO: Pod downwardapi-volume-d6ca916e-05cd-411a-aed0-4f1f0bd3ea6b no longer exists + [AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 - Dec 14 16:30:16.324: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "events-7375" for this suite. 12/14/22 16:30:16.331 + Feb 25 10:30:42.567: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-5158" for this suite. 02/25/23 10:30:42.648 << End Captured GinkgoWriter Output ------------------------------ -S +SSSSS ------------------------------ -[sig-network] Proxy version v1 - A set of valid responses are returned for both pod and service Proxy [Conformance] - test/e2e/network/proxy.go:380 -[BeforeEach] version v1 +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + listing mutating webhooks should work [Conformance] + test/e2e/apimachinery/webhook.go:655 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:30:16.344 -Dec 14 16:30:16.344: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename proxy 12/14/22 16:30:16.347 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:16.374 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:16.377 -[It] A set of valid responses are returned for both pod and service Proxy [Conformance] - test/e2e/network/proxy.go:380 -Dec 14 16:30:16.380: INFO: Creating pod... -Dec 14 16:30:16.413: INFO: Waiting up to 5m0s for pod "agnhost" in namespace "proxy-9238" to be "running" -Dec 14 16:30:16.421: INFO: Pod "agnhost": Phase="Pending", Reason="", readiness=false. Elapsed: 7.186133ms -Dec 14 16:30:18.432: INFO: Pod "agnhost": Phase="Running", Reason="", readiness=true. Elapsed: 2.018660456s -Dec 14 16:30:18.432: INFO: Pod "agnhost" satisfied condition "running" -Dec 14 16:30:18.432: INFO: Creating service... -Dec 14 16:30:18.455: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/pods/agnhost/proxy?method=DELETE -Dec 14 16:30:18.469: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE -Dec 14 16:30:18.469: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/pods/agnhost/proxy?method=OPTIONS -Dec 14 16:30:18.474: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS -Dec 14 16:30:18.474: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/pods/agnhost/proxy?method=PATCH -Dec 14 16:30:18.481: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH -Dec 14 16:30:18.481: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/pods/agnhost/proxy?method=POST -Dec 14 16:30:18.489: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST -Dec 14 16:30:18.490: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/pods/agnhost/proxy?method=PUT -Dec 14 16:30:18.494: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT -Dec 14 16:30:18.494: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/services/e2e-proxy-test-service/proxy?method=DELETE -Dec 14 16:30:18.501: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE -Dec 14 16:30:18.501: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/services/e2e-proxy-test-service/proxy?method=OPTIONS -Dec 14 16:30:18.510: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS -Dec 14 16:30:18.510: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/services/e2e-proxy-test-service/proxy?method=PATCH -Dec 14 16:30:18.518: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH -Dec 14 16:30:18.518: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/services/e2e-proxy-test-service/proxy?method=POST -Dec 14 16:30:18.523: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST -Dec 14 16:30:18.523: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/services/e2e-proxy-test-service/proxy?method=PUT -Dec 14 16:30:18.532: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT -Dec 14 16:30:18.532: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/pods/agnhost/proxy?method=GET -Dec 14 16:30:18.535: INFO: http.Client request:GET StatusCode:301 -Dec 14 16:30:18.535: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/services/e2e-proxy-test-service/proxy?method=GET -Dec 14 16:30:18.547: INFO: http.Client request:GET StatusCode:301 -Dec 14 16:30:18.547: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/pods/agnhost/proxy?method=HEAD -Dec 14 16:30:18.550: INFO: http.Client request:HEAD StatusCode:301 -Dec 14 16:30:18.550: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/services/e2e-proxy-test-service/proxy?method=HEAD -Dec 14 16:30:18.561: INFO: http.Client request:HEAD StatusCode:301 -[AfterEach] version v1 +STEP: Creating a kubernetes client 02/25/23 10:30:42.676 +Feb 25 10:30:42.676: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 10:30:42.679 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:42.71 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:42.716 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 +STEP: Setting up server cert 02/25/23 10:30:42.76 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:30:43.697 +STEP: Deploying the webhook pod 02/25/23 10:30:43.724 +STEP: Wait for the deployment to be ready 02/25/23 10:30:43.763 +Feb 25 10:30:43.779: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created +Feb 25 10:30:45.803: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 30, 43, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 30, 43, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 30, 43, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 30, 43, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} +STEP: Deploying the webhook service 02/25/23 10:30:47.81 +STEP: Verifying the service has paired with the endpoint 02/25/23 10:30:47.839 +Feb 25 10:30:48.840: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] listing mutating webhooks should work [Conformance] + test/e2e/apimachinery/webhook.go:655 +STEP: Listing all of the created validation webhooks 02/25/23 10:30:49.142 +STEP: Creating a configMap that should be mutated 02/25/23 10:30:49.302 +STEP: Deleting the collection of validation webhooks 02/25/23 10:30:49.387 +STEP: Creating a configMap that should not be mutated 02/25/23 10:30:49.62 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:30:18.562: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "proxy-9238" for this suite. 12/14/22 16:30:18.568 -{"msg":"PASSED [sig-network] Proxy version v1 A set of valid responses are returned for both pod and service Proxy [Conformance]","completed":151,"skipped":2753,"failed":0} +Feb 25 10:30:49.675: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-7125" for this suite. 02/25/23 10:30:49.703 +STEP: Destroying namespace "webhook-7125-markers" for this suite. 02/25/23 10:30:49.731 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing mutating webhooks should work [Conformance]","completed":143,"skipped":2612,"failed":0} ------------------------------ -• [2.232 seconds] -[sig-network] Proxy -test/e2e/network/common/framework.go:23 - version v1 - test/e2e/network/proxy.go:74 - A set of valid responses are returned for both pod and service Proxy [Conformance] - test/e2e/network/proxy.go:380 +• [SLOW TEST] [7.223 seconds] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + listing mutating webhooks should work [Conformance] + test/e2e/apimachinery/webhook.go:655 Begin Captured GinkgoWriter Output >> - [BeforeEach] version v1 + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:30:16.344 - Dec 14 16:30:16.344: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename proxy 12/14/22 16:30:16.347 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:16.374 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:16.377 - [It] A set of valid responses are returned for both pod and service Proxy [Conformance] - test/e2e/network/proxy.go:380 - Dec 14 16:30:16.380: INFO: Creating pod... - Dec 14 16:30:16.413: INFO: Waiting up to 5m0s for pod "agnhost" in namespace "proxy-9238" to be "running" - Dec 14 16:30:16.421: INFO: Pod "agnhost": Phase="Pending", Reason="", readiness=false. Elapsed: 7.186133ms - Dec 14 16:30:18.432: INFO: Pod "agnhost": Phase="Running", Reason="", readiness=true. Elapsed: 2.018660456s - Dec 14 16:30:18.432: INFO: Pod "agnhost" satisfied condition "running" - Dec 14 16:30:18.432: INFO: Creating service... - Dec 14 16:30:18.455: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/pods/agnhost/proxy?method=DELETE - Dec 14 16:30:18.469: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE - Dec 14 16:30:18.469: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/pods/agnhost/proxy?method=OPTIONS - Dec 14 16:30:18.474: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS - Dec 14 16:30:18.474: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/pods/agnhost/proxy?method=PATCH - Dec 14 16:30:18.481: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH - Dec 14 16:30:18.481: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/pods/agnhost/proxy?method=POST - Dec 14 16:30:18.489: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST - Dec 14 16:30:18.490: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/pods/agnhost/proxy?method=PUT - Dec 14 16:30:18.494: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT - Dec 14 16:30:18.494: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/services/e2e-proxy-test-service/proxy?method=DELETE - Dec 14 16:30:18.501: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE - Dec 14 16:30:18.501: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/services/e2e-proxy-test-service/proxy?method=OPTIONS - Dec 14 16:30:18.510: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS - Dec 14 16:30:18.510: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/services/e2e-proxy-test-service/proxy?method=PATCH - Dec 14 16:30:18.518: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH - Dec 14 16:30:18.518: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/services/e2e-proxy-test-service/proxy?method=POST - Dec 14 16:30:18.523: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST - Dec 14 16:30:18.523: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/services/e2e-proxy-test-service/proxy?method=PUT - Dec 14 16:30:18.532: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT - Dec 14 16:30:18.532: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/pods/agnhost/proxy?method=GET - Dec 14 16:30:18.535: INFO: http.Client request:GET StatusCode:301 - Dec 14 16:30:18.535: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/services/e2e-proxy-test-service/proxy?method=GET - Dec 14 16:30:18.547: INFO: http.Client request:GET StatusCode:301 - Dec 14 16:30:18.547: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/pods/agnhost/proxy?method=HEAD - Dec 14 16:30:18.550: INFO: http.Client request:HEAD StatusCode:301 - Dec 14 16:30:18.550: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-9238/services/e2e-proxy-test-service/proxy?method=HEAD - Dec 14 16:30:18.561: INFO: http.Client request:HEAD StatusCode:301 - [AfterEach] version v1 + STEP: Creating a kubernetes client 02/25/23 10:30:42.676 + Feb 25 10:30:42.676: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 10:30:42.679 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:42.71 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:42.716 + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 + STEP: Setting up server cert 02/25/23 10:30:42.76 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:30:43.697 + STEP: Deploying the webhook pod 02/25/23 10:30:43.724 + STEP: Wait for the deployment to be ready 02/25/23 10:30:43.763 + Feb 25 10:30:43.779: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created + Feb 25 10:30:45.803: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 30, 43, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 30, 43, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 30, 43, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 30, 43, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} + STEP: Deploying the webhook service 02/25/23 10:30:47.81 + STEP: Verifying the service has paired with the endpoint 02/25/23 10:30:47.839 + Feb 25 10:30:48.840: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] listing mutating webhooks should work [Conformance] + test/e2e/apimachinery/webhook.go:655 + STEP: Listing all of the created validation webhooks 02/25/23 10:30:49.142 + STEP: Creating a configMap that should be mutated 02/25/23 10:30:49.302 + STEP: Deleting the collection of validation webhooks 02/25/23 10:30:49.387 + STEP: Creating a configMap that should not be mutated 02/25/23 10:30:49.62 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:30:18.562: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "proxy-9238" for this suite. 12/14/22 16:30:18.568 + Feb 25 10:30:49.675: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-7125" for this suite. 02/25/23 10:30:49.703 + STEP: Destroying namespace "webhook-7125-markers" for this suite. 02/25/23 10:30:49.731 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance] test/e2e/apps/daemon_set.go:193 [BeforeEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:30:18.579 -Dec 14 16:30:18.579: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename daemonsets 12/14/22 16:30:18.583 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:18.604 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:18.607 +STEP: Creating a kubernetes client 02/25/23 10:30:49.919 +Feb 25 10:30:49.920: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename daemonsets 02/25/23 10:30:49.923 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:49.98 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:49.987 [BeforeEach] [sig-apps] Daemon set [Serial] test/e2e/apps/daemon_set.go:145 [It] should run and stop complex daemon [Conformance] test/e2e/apps/daemon_set.go:193 -Dec 14 16:30:18.652: INFO: Creating daemon "daemon-set" with a node selector -STEP: Initially, daemon pods should not be running on any nodes. 12/14/22 16:30:18.659 -Dec 14 16:30:18.664: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:30:18.665: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set -STEP: Change node label to blue, check that daemon pod is launched. 12/14/22 16:30:18.665 -Dec 14 16:30:18.701: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:30:18.701: INFO: Node kiedai4iakae-3 is running 0 daemon pod, expected 1 -Dec 14 16:30:19.720: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 -Dec 14 16:30:19.720: INFO: Number of running nodes: 1, number of available pods: 1 in daemonset daemon-set -STEP: Update the node label to green, and wait for daemons to be unscheduled 12/14/22 16:30:19.727 -Dec 14 16:30:19.776: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 -Dec 14 16:30:19.776: INFO: Number of running nodes: 0, number of available pods: 1 in daemonset daemon-set -Dec 14 16:30:20.791: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:30:20.791: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set -STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate 12/14/22 16:30:20.791 -Dec 14 16:30:20.820: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:30:20.820: INFO: Node kiedai4iakae-3 is running 0 daemon pod, expected 1 -Dec 14 16:30:21.827: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:30:21.827: INFO: Node kiedai4iakae-3 is running 0 daemon pod, expected 1 -Dec 14 16:30:22.829: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:30:22.830: INFO: Node kiedai4iakae-3 is running 0 daemon pod, expected 1 -Dec 14 16:30:23.832: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 -Dec 14 16:30:23.832: INFO: Number of running nodes: 1, number of available pods: 1 in daemonset daemon-set +Feb 25 10:30:50.063: INFO: Creating daemon "daemon-set" with a node selector +STEP: Initially, daemon pods should not be running on any nodes. 02/25/23 10:30:50.109 +Feb 25 10:30:50.128: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:30:50.128: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set +STEP: Change node label to blue, check that daemon pod is launched. 02/25/23 10:30:50.128 +Feb 25 10:30:50.280: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:30:50.280: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 +Feb 25 10:30:51.509: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:30:51.509: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 +Feb 25 10:30:52.982: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:30:53.300: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 +Feb 25 10:30:53.312: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:30:53.312: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 +Feb 25 10:30:54.290: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:30:54.290: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 +Feb 25 10:30:55.303: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 +Feb 25 10:30:55.303: INFO: Number of running nodes: 1, number of available pods: 1 in daemonset daemon-set +STEP: Update the node label to green, and wait for daemons to be unscheduled 02/25/23 10:30:55.319 +Feb 25 10:30:55.457: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 +Feb 25 10:30:55.465: INFO: Number of running nodes: 0, number of available pods: 1 in daemonset daemon-set +Feb 25 10:30:56.520: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:30:56.520: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set +STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate 02/25/23 10:30:56.52 +Feb 25 10:30:57.088: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:30:57.107: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 +Feb 25 10:30:58.865: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:30:58.865: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 +Feb 25 10:30:59.863: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:30:59.863: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 +Feb 25 10:31:00.953: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:31:00.954: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 +Feb 25 10:31:01.863: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 +Feb 25 10:31:01.863: INFO: Number of running nodes: 1, number of available pods: 1 in daemonset daemon-set [AfterEach] [sig-apps] Daemon set [Serial] test/e2e/apps/daemon_set.go:110 -STEP: Deleting DaemonSet "daemon-set" 12/14/22 16:30:23.848 -STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-1582, will wait for the garbage collector to delete the pods 12/14/22 16:30:23.848 -Dec 14 16:30:23.917: INFO: Deleting DaemonSet.extensions daemon-set took: 9.445748ms -Dec 14 16:30:24.019: INFO: Terminating DaemonSet.extensions daemon-set pods took: 101.426902ms -Dec 14 16:30:26.630: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:30:26.630: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set -Dec 14 16:30:26.636: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"17872"},"items":null} +STEP: Deleting DaemonSet "daemon-set" 02/25/23 10:31:01.879 +STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-9802, will wait for the garbage collector to delete the pods 02/25/23 10:31:01.879 +Feb 25 10:31:01.963: INFO: Deleting DaemonSet.extensions daemon-set took: 23.401989ms +Feb 25 10:31:02.064: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.776278ms +Feb 25 10:31:04.772: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:31:04.772: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set +Feb 25 10:31:04.780: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"23236"},"items":null} -Dec 14 16:30:26.640: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"17872"},"items":null} +Feb 25 10:31:04.786: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"23236"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:187 -Dec 14 16:30:26.675: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "daemonsets-1582" for this suite. 12/14/22 16:30:26.688 -{"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance]","completed":152,"skipped":2767,"failed":0} +Feb 25 10:31:04.838: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "daemonsets-9802" for this suite. 02/25/23 10:31:04.904 +{"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop complex daemon [Conformance]","completed":144,"skipped":2633,"failed":0} ------------------------------ -• [SLOW TEST] [8.120 seconds] +• [SLOW TEST] [15.092 seconds] [sig-apps] Daemon set [Serial] test/e2e/apps/framework.go:23 should run and stop complex daemon [Conformance] @@ -14073,93 +13171,104 @@ test/e2e/apps/framework.go:23 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:30:18.579 - Dec 14 16:30:18.579: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename daemonsets 12/14/22 16:30:18.583 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:18.604 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:18.607 + STEP: Creating a kubernetes client 02/25/23 10:30:49.919 + Feb 25 10:30:49.920: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename daemonsets 02/25/23 10:30:49.923 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:30:49.98 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:30:49.987 [BeforeEach] [sig-apps] Daemon set [Serial] test/e2e/apps/daemon_set.go:145 [It] should run and stop complex daemon [Conformance] test/e2e/apps/daemon_set.go:193 - Dec 14 16:30:18.652: INFO: Creating daemon "daemon-set" with a node selector - STEP: Initially, daemon pods should not be running on any nodes. 12/14/22 16:30:18.659 - Dec 14 16:30:18.664: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:30:18.665: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set - STEP: Change node label to blue, check that daemon pod is launched. 12/14/22 16:30:18.665 - Dec 14 16:30:18.701: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:30:18.701: INFO: Node kiedai4iakae-3 is running 0 daemon pod, expected 1 - Dec 14 16:30:19.720: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 - Dec 14 16:30:19.720: INFO: Number of running nodes: 1, number of available pods: 1 in daemonset daemon-set - STEP: Update the node label to green, and wait for daemons to be unscheduled 12/14/22 16:30:19.727 - Dec 14 16:30:19.776: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 - Dec 14 16:30:19.776: INFO: Number of running nodes: 0, number of available pods: 1 in daemonset daemon-set - Dec 14 16:30:20.791: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:30:20.791: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set - STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate 12/14/22 16:30:20.791 - Dec 14 16:30:20.820: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:30:20.820: INFO: Node kiedai4iakae-3 is running 0 daemon pod, expected 1 - Dec 14 16:30:21.827: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:30:21.827: INFO: Node kiedai4iakae-3 is running 0 daemon pod, expected 1 - Dec 14 16:30:22.829: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:30:22.830: INFO: Node kiedai4iakae-3 is running 0 daemon pod, expected 1 - Dec 14 16:30:23.832: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 - Dec 14 16:30:23.832: INFO: Number of running nodes: 1, number of available pods: 1 in daemonset daemon-set + Feb 25 10:30:50.063: INFO: Creating daemon "daemon-set" with a node selector + STEP: Initially, daemon pods should not be running on any nodes. 02/25/23 10:30:50.109 + Feb 25 10:30:50.128: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:30:50.128: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set + STEP: Change node label to blue, check that daemon pod is launched. 02/25/23 10:30:50.128 + Feb 25 10:30:50.280: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:30:50.280: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 + Feb 25 10:30:51.509: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:30:51.509: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 + Feb 25 10:30:52.982: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:30:53.300: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 + Feb 25 10:30:53.312: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:30:53.312: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 + Feb 25 10:30:54.290: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:30:54.290: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 + Feb 25 10:30:55.303: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 + Feb 25 10:30:55.303: INFO: Number of running nodes: 1, number of available pods: 1 in daemonset daemon-set + STEP: Update the node label to green, and wait for daemons to be unscheduled 02/25/23 10:30:55.319 + Feb 25 10:30:55.457: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 + Feb 25 10:30:55.465: INFO: Number of running nodes: 0, number of available pods: 1 in daemonset daemon-set + Feb 25 10:30:56.520: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:30:56.520: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set + STEP: Update DaemonSet node selector to green, and change its update strategy to RollingUpdate 02/25/23 10:30:56.52 + Feb 25 10:30:57.088: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:30:57.107: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 + Feb 25 10:30:58.865: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:30:58.865: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 + Feb 25 10:30:59.863: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:30:59.863: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 + Feb 25 10:31:00.953: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:31:00.954: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 + Feb 25 10:31:01.863: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 + Feb 25 10:31:01.863: INFO: Number of running nodes: 1, number of available pods: 1 in daemonset daemon-set [AfterEach] [sig-apps] Daemon set [Serial] test/e2e/apps/daemon_set.go:110 - STEP: Deleting DaemonSet "daemon-set" 12/14/22 16:30:23.848 - STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-1582, will wait for the garbage collector to delete the pods 12/14/22 16:30:23.848 - Dec 14 16:30:23.917: INFO: Deleting DaemonSet.extensions daemon-set took: 9.445748ms - Dec 14 16:30:24.019: INFO: Terminating DaemonSet.extensions daemon-set pods took: 101.426902ms - Dec 14 16:30:26.630: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:30:26.630: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set - Dec 14 16:30:26.636: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"17872"},"items":null} + STEP: Deleting DaemonSet "daemon-set" 02/25/23 10:31:01.879 + STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-9802, will wait for the garbage collector to delete the pods 02/25/23 10:31:01.879 + Feb 25 10:31:01.963: INFO: Deleting DaemonSet.extensions daemon-set took: 23.401989ms + Feb 25 10:31:02.064: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.776278ms + Feb 25 10:31:04.772: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:31:04.772: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set + Feb 25 10:31:04.780: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"23236"},"items":null} - Dec 14 16:30:26.640: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"17872"},"items":null} + Feb 25 10:31:04.786: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"23236"},"items":null} [AfterEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:187 - Dec 14 16:30:26.675: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "daemonsets-1582" for this suite. 12/14/22 16:30:26.688 + Feb 25 10:31:04.838: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "daemonsets-9802" for this suite. 02/25/23 10:31:04.904 << End Captured GinkgoWriter Output ------------------------------ -S +SSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance] test/e2e/apimachinery/webhook.go:307 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:30:26.7 -Dec 14 16:30:26.700: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 16:30:26.702 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:26.731 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:26.735 +STEP: Creating a kubernetes client 02/25/23 10:31:05.018 +Feb 25 10:31:05.079: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 10:31:05.084 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:31:05.184 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:31:05.203 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 16:30:26.761 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:30:27.444 -STEP: Deploying the webhook pod 12/14/22 16:30:27.463 -STEP: Wait for the deployment to be ready 12/14/22 16:30:27.48 -Dec 14 16:30:27.487: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created -STEP: Deploying the webhook service 12/14/22 16:30:29.507 -STEP: Verifying the service has paired with the endpoint 12/14/22 16:30:29.532 -Dec 14 16:30:30.534: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +STEP: Setting up server cert 02/25/23 10:31:05.239 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:31:07.719 +STEP: Deploying the webhook pod 02/25/23 10:31:07.732 +STEP: Wait for the deployment to be ready 02/25/23 10:31:07.759 +Feb 25 10:31:07.774: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created +Feb 25 10:31:09.797: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 31, 7, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 31, 7, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 31, 7, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 31, 7, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} +STEP: Deploying the webhook service 02/25/23 10:31:11.804 +STEP: Verifying the service has paired with the endpoint 02/25/23 10:31:11.827 +Feb 25 10:31:12.828: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should deny crd creation [Conformance] test/e2e/apimachinery/webhook.go:307 -STEP: Registering the crd webhook via the AdmissionRegistration API 12/14/22 16:30:30.542 -STEP: Creating a custom resource definition that should be denied by the webhook 12/14/22 16:30:30.573 -Dec 14 16:30:30.574: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 +STEP: Registering the crd webhook via the AdmissionRegistration API 02/25/23 10:31:12.84 +STEP: Creating a custom resource definition that should be denied by the webhook 02/25/23 10:31:12.887 +Feb 25 10:31:12.887: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:30:30.600: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-8806" for this suite. 12/14/22 16:30:30.606 -STEP: Destroying namespace "webhook-8806-markers" for this suite. 12/14/22 16:30:30.615 +Feb 25 10:31:12.971: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-8740" for this suite. 02/25/23 10:31:12.988 +STEP: Destroying namespace "webhook-8740-markers" for this suite. 02/25/23 10:31:13.012 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance]","completed":153,"skipped":2768,"failed":0} +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should deny crd creation [Conformance]","completed":145,"skipped":2636,"failed":0} ------------------------------ -• [3.971 seconds] +• [SLOW TEST] [8.411 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/framework.go:23 should deny crd creation [Conformance] @@ -14168,4261 +13277,4572 @@ test/e2e/apimachinery/framework.go:23 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:30:26.7 - Dec 14 16:30:26.700: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 16:30:26.702 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:26.731 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:26.735 + STEP: Creating a kubernetes client 02/25/23 10:31:05.018 + Feb 25 10:31:05.079: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 10:31:05.084 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:31:05.184 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:31:05.203 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 16:30:26.761 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:30:27.444 - STEP: Deploying the webhook pod 12/14/22 16:30:27.463 - STEP: Wait for the deployment to be ready 12/14/22 16:30:27.48 - Dec 14 16:30:27.487: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created - STEP: Deploying the webhook service 12/14/22 16:30:29.507 - STEP: Verifying the service has paired with the endpoint 12/14/22 16:30:29.532 - Dec 14 16:30:30.534: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + STEP: Setting up server cert 02/25/23 10:31:05.239 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:31:07.719 + STEP: Deploying the webhook pod 02/25/23 10:31:07.732 + STEP: Wait for the deployment to be ready 02/25/23 10:31:07.759 + Feb 25 10:31:07.774: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created + Feb 25 10:31:09.797: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 31, 7, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 31, 7, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 31, 7, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 31, 7, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} + STEP: Deploying the webhook service 02/25/23 10:31:11.804 + STEP: Verifying the service has paired with the endpoint 02/25/23 10:31:11.827 + Feb 25 10:31:12.828: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 [It] should deny crd creation [Conformance] test/e2e/apimachinery/webhook.go:307 - STEP: Registering the crd webhook via the AdmissionRegistration API 12/14/22 16:30:30.542 - STEP: Creating a custom resource definition that should be denied by the webhook 12/14/22 16:30:30.573 - Dec 14 16:30:30.574: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 + STEP: Registering the crd webhook via the AdmissionRegistration API 02/25/23 10:31:12.84 + STEP: Creating a custom resource definition that should be denied by the webhook 02/25/23 10:31:12.887 + Feb 25 10:31:12.887: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:30:30.600: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-8806" for this suite. 12/14/22 16:30:30.606 - STEP: Destroying namespace "webhook-8806-markers" for this suite. 12/14/22 16:30:30.615 + Feb 25 10:31:12.971: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-8740" for this suite. 02/25/23 10:31:12.988 + STEP: Destroying namespace "webhook-8740-markers" for this suite. 02/25/23 10:31:13.012 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:104 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSS +S +------------------------------ +[sig-apps] Daemon set [Serial] + should run and stop simple daemon [Conformance] + test/e2e/apps/daemon_set.go:165 +[BeforeEach] [sig-apps] Daemon set [Serial] + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:31:13.432 +Feb 25 10:31:13.432: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename daemonsets 02/25/23 10:31:13.435 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:31:13.48 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:31:13.488 +[BeforeEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:145 +[It] should run and stop simple daemon [Conformance] + test/e2e/apps/daemon_set.go:165 +STEP: Creating simple DaemonSet "daemon-set" 02/25/23 10:31:13.572 +STEP: Check that daemon pods launch on every node of the cluster. 02/25/23 10:31:13.594 +Feb 25 10:31:13.628: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:31:13.628: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 10:31:14.675: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:31:14.675: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 10:31:15.656: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:31:15.656: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 10:31:16.650: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 10:31:16.650: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 +Feb 25 10:31:17.653: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 +Feb 25 10:31:17.653: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set +STEP: Stop a daemon pod, check that the daemon pod is revived. 02/25/23 10:31:17.66 +Feb 25 10:31:17.722: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 10:31:17.722: INFO: Node diek3dua9gay-2 is running 0 daemon pod, expected 1 +Feb 25 10:31:18.749: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 10:31:18.749: INFO: Node diek3dua9gay-2 is running 0 daemon pod, expected 1 +Feb 25 10:31:19.736: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 10:31:19.737: INFO: Node diek3dua9gay-2 is running 0 daemon pod, expected 1 +Feb 25 10:31:20.744: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 10:31:20.744: INFO: Node diek3dua9gay-2 is running 0 daemon pod, expected 1 +Feb 25 10:31:21.927: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 +Feb 25 10:31:21.928: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set +[AfterEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:110 +STEP: Deleting DaemonSet "daemon-set" 02/25/23 10:31:22.015 +STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-1198, will wait for the garbage collector to delete the pods 02/25/23 10:31:22.015 +Feb 25 10:31:22.157: INFO: Deleting DaemonSet.extensions daemon-set took: 35.868167ms +Feb 25 10:31:22.257: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.541689ms +Feb 25 10:31:24.265: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:31:24.265: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set +Feb 25 10:31:24.273: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"23445"},"items":null} + +Feb 25 10:31:24.281: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"23445"},"items":null} + +[AfterEach] [sig-apps] Daemon set [Serial] + test/e2e/framework/framework.go:187 +Feb 25 10:31:24.313: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "daemonsets-1198" for this suite. 02/25/23 10:31:24.321 +{"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop simple daemon [Conformance]","completed":146,"skipped":2637,"failed":0} +------------------------------ +• [SLOW TEST] [10.905 seconds] +[sig-apps] Daemon set [Serial] +test/e2e/apps/framework.go:23 + should run and stop simple daemon [Conformance] + test/e2e/apps/daemon_set.go:165 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-apps] Daemon set [Serial] + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:31:13.432 + Feb 25 10:31:13.432: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename daemonsets 02/25/23 10:31:13.435 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:31:13.48 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:31:13.488 + [BeforeEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:145 + [It] should run and stop simple daemon [Conformance] + test/e2e/apps/daemon_set.go:165 + STEP: Creating simple DaemonSet "daemon-set" 02/25/23 10:31:13.572 + STEP: Check that daemon pods launch on every node of the cluster. 02/25/23 10:31:13.594 + Feb 25 10:31:13.628: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:31:13.628: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 10:31:14.675: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:31:14.675: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 10:31:15.656: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:31:15.656: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 10:31:16.650: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 10:31:16.650: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 + Feb 25 10:31:17.653: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 + Feb 25 10:31:17.653: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set + STEP: Stop a daemon pod, check that the daemon pod is revived. 02/25/23 10:31:17.66 + Feb 25 10:31:17.722: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 10:31:17.722: INFO: Node diek3dua9gay-2 is running 0 daemon pod, expected 1 + Feb 25 10:31:18.749: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 10:31:18.749: INFO: Node diek3dua9gay-2 is running 0 daemon pod, expected 1 + Feb 25 10:31:19.736: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 10:31:19.737: INFO: Node diek3dua9gay-2 is running 0 daemon pod, expected 1 + Feb 25 10:31:20.744: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 10:31:20.744: INFO: Node diek3dua9gay-2 is running 0 daemon pod, expected 1 + Feb 25 10:31:21.927: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 + Feb 25 10:31:21.928: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set + [AfterEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:110 + STEP: Deleting DaemonSet "daemon-set" 02/25/23 10:31:22.015 + STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-1198, will wait for the garbage collector to delete the pods 02/25/23 10:31:22.015 + Feb 25 10:31:22.157: INFO: Deleting DaemonSet.extensions daemon-set took: 35.868167ms + Feb 25 10:31:22.257: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.541689ms + Feb 25 10:31:24.265: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:31:24.265: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set + Feb 25 10:31:24.273: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"23445"},"items":null} + + Feb 25 10:31:24.281: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"23445"},"items":null} + + [AfterEach] [sig-apps] Daemon set [Serial] + test/e2e/framework/framework.go:187 + Feb 25 10:31:24.313: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "daemonsets-1198" for this suite. 02/25/23 10:31:24.321 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-node] Secrets + should be consumable via the environment [NodeConformance] [Conformance] + test/e2e/common/node/secrets.go:94 +[BeforeEach] [sig-node] Secrets + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:31:24.341 +Feb 25 10:31:24.341: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename secrets 02/25/23 10:31:24.346 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:31:24.388 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:31:24.397 +[It] should be consumable via the environment [NodeConformance] [Conformance] + test/e2e/common/node/secrets.go:94 +STEP: creating secret secrets-199/secret-test-cd3966f6-af79-4f88-b70c-b72963e3e8c2 02/25/23 10:31:24.405 +STEP: Creating a pod to test consume secrets 02/25/23 10:31:24.425 +Feb 25 10:31:24.444: INFO: Waiting up to 5m0s for pod "pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c" in namespace "secrets-199" to be "Succeeded or Failed" +Feb 25 10:31:24.453: INFO: Pod "pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c": Phase="Pending", Reason="", readiness=false. Elapsed: 8.628669ms +Feb 25 10:31:26.467: INFO: Pod "pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022362637s +Feb 25 10:31:28.471: INFO: Pod "pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.025965758s +Feb 25 10:31:30.462: INFO: Pod "pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.017521243s +STEP: Saw pod success 02/25/23 10:31:30.462 +Feb 25 10:31:30.463: INFO: Pod "pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c" satisfied condition "Succeeded or Failed" +Feb 25 10:31:30.473: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c container env-test: +STEP: delete the pod 02/25/23 10:31:30.492 +Feb 25 10:31:30.518: INFO: Waiting for pod pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c to disappear +Feb 25 10:31:30.528: INFO: Pod pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c no longer exists +[AfterEach] [sig-node] Secrets + test/e2e/framework/framework.go:187 +Feb 25 10:31:30.529: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-199" for this suite. 02/25/23 10:31:30.555 +{"msg":"PASSED [sig-node] Secrets should be consumable via the environment [NodeConformance] [Conformance]","completed":147,"skipped":2662,"failed":0} +------------------------------ +• [SLOW TEST] [6.227 seconds] +[sig-node] Secrets +test/e2e/common/node/framework.go:23 + should be consumable via the environment [NodeConformance] [Conformance] + test/e2e/common/node/secrets.go:94 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-node] Secrets + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:31:24.341 + Feb 25 10:31:24.341: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename secrets 02/25/23 10:31:24.346 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:31:24.388 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:31:24.397 + [It] should be consumable via the environment [NodeConformance] [Conformance] + test/e2e/common/node/secrets.go:94 + STEP: creating secret secrets-199/secret-test-cd3966f6-af79-4f88-b70c-b72963e3e8c2 02/25/23 10:31:24.405 + STEP: Creating a pod to test consume secrets 02/25/23 10:31:24.425 + Feb 25 10:31:24.444: INFO: Waiting up to 5m0s for pod "pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c" in namespace "secrets-199" to be "Succeeded or Failed" + Feb 25 10:31:24.453: INFO: Pod "pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c": Phase="Pending", Reason="", readiness=false. Elapsed: 8.628669ms + Feb 25 10:31:26.467: INFO: Pod "pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022362637s + Feb 25 10:31:28.471: INFO: Pod "pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.025965758s + Feb 25 10:31:30.462: INFO: Pod "pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.017521243s + STEP: Saw pod success 02/25/23 10:31:30.462 + Feb 25 10:31:30.463: INFO: Pod "pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c" satisfied condition "Succeeded or Failed" + Feb 25 10:31:30.473: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c container env-test: + STEP: delete the pod 02/25/23 10:31:30.492 + Feb 25 10:31:30.518: INFO: Waiting for pod pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c to disappear + Feb 25 10:31:30.528: INFO: Pod pod-configmaps-a1135f4a-92cf-46ec-abb7-22ebc7cd1e1c no longer exists + [AfterEach] [sig-node] Secrets + test/e2e/framework/framework.go:187 + Feb 25 10:31:30.529: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "secrets-199" for this suite. 02/25/23 10:31:30.555 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSS ------------------------------ -[sig-apps] DisruptionController - should block an eviction until the PDB is updated to allow it [Conformance] - test/e2e/apps/disruption.go:346 -[BeforeEach] [sig-apps] DisruptionController +[sig-network] Services + should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance] + test/e2e/network/service.go:2189 +[BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:30:30.672 -Dec 14 16:30:30.672: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename disruption 12/14/22 16:30:30.674 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:30.698 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:30.702 -[BeforeEach] [sig-apps] DisruptionController - test/e2e/apps/disruption.go:71 -[It] should block an eviction until the PDB is updated to allow it [Conformance] - test/e2e/apps/disruption.go:346 -STEP: Creating a pdb that targets all three pods in a test replica set 12/14/22 16:30:30.705 -STEP: Waiting for the pdb to be processed 12/14/22 16:30:30.712 -STEP: First trying to evict a pod which shouldn't be evictable 12/14/22 16:30:32.733 -STEP: Waiting for all pods to be running 12/14/22 16:30:32.733 -Dec 14 16:30:32.740: INFO: pods: 0 < 3 -STEP: locating a running pod 12/14/22 16:30:34.749 -STEP: Updating the pdb to allow a pod to be evicted 12/14/22 16:30:34.768 -STEP: Waiting for the pdb to be processed 12/14/22 16:30:34.781 -STEP: Trying to evict the same pod we tried earlier which should now be evictable 12/14/22 16:30:36.797 -STEP: Waiting for all pods to be running 12/14/22 16:30:36.797 -STEP: Waiting for the pdb to observed all healthy pods 12/14/22 16:30:36.806 -STEP: Patching the pdb to disallow a pod to be evicted 12/14/22 16:30:36.838 -STEP: Waiting for the pdb to be processed 12/14/22 16:30:36.875 -STEP: Waiting for all pods to be running 12/14/22 16:30:38.914 -STEP: locating a running pod 12/14/22 16:30:38.922 -STEP: Deleting the pdb to allow a pod to be evicted 12/14/22 16:30:38.939 -STEP: Waiting for the pdb to be deleted 12/14/22 16:30:38.948 -STEP: Trying to evict the same pod we tried earlier which should now be evictable 12/14/22 16:30:38.952 -STEP: Waiting for all pods to be running 12/14/22 16:30:38.952 -[AfterEach] [sig-apps] DisruptionController +STEP: Creating a kubernetes client 02/25/23 10:31:30.578 +Feb 25 10:31:30.579: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 10:31:30.593 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:31:30.636 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:31:30.64 +[BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 +[It] should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance] + test/e2e/network/service.go:2189 +STEP: creating service in namespace services-8674 02/25/23 10:31:30.646 +STEP: creating service affinity-clusterip-transition in namespace services-8674 02/25/23 10:31:30.647 +STEP: creating replication controller affinity-clusterip-transition in namespace services-8674 02/25/23 10:31:30.685 +I0225 10:31:30.717266 13 runners.go:193] Created replication controller with name: affinity-clusterip-transition, namespace: services-8674, replica count: 3 +I0225 10:31:33.768298 13 runners.go:193] affinity-clusterip-transition Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Feb 25 10:31:33.785: INFO: Creating new exec pod +Feb 25 10:31:33.812: INFO: Waiting up to 5m0s for pod "execpod-affinityfp4mz" in namespace "services-8674" to be "running" +Feb 25 10:31:33.833: INFO: Pod "execpod-affinityfp4mz": Phase="Pending", Reason="", readiness=false. Elapsed: 21.292329ms +Feb 25 10:31:35.841: INFO: Pod "execpod-affinityfp4mz": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02892199s +Feb 25 10:31:37.844: INFO: Pod "execpod-affinityfp4mz": Phase="Running", Reason="", readiness=true. Elapsed: 4.031626635s +Feb 25 10:31:37.844: INFO: Pod "execpod-affinityfp4mz" satisfied condition "running" +Feb 25 10:31:38.845: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-8674 exec execpod-affinityfp4mz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-clusterip-transition 80' +Feb 25 10:31:39.289: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-clusterip-transition 80\nConnection to affinity-clusterip-transition 80 port [tcp/http] succeeded!\n" +Feb 25 10:31:39.289: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 10:31:39.289: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-8674 exec execpod-affinityfp4mz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.12.194 80' +Feb 25 10:31:39.555: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.12.194 80\nConnection to 10.233.12.194 80 port [tcp/http] succeeded!\n" +Feb 25 10:31:39.555: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 10:31:39.582: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-8674 exec execpod-affinityfp4mz -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.12.194:80/ ; done' +Feb 25 10:31:40.521: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n" +Feb 25 10:31:40.522: INFO: stdout: "\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-gpgkh\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-gpgkh\naffinity-clusterip-transition-gpgkh\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-q7c8w\naffinity-clusterip-transition-q7c8w\naffinity-clusterip-transition-gpgkh\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-q7c8w\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p" +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-gpgkh +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-gpgkh +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-gpgkh +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-q7c8w +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-q7c8w +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-gpgkh +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-q7c8w +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:40.595: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-8674 exec execpod-affinityfp4mz -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.12.194:80/ ; done' +Feb 25 10:31:41.384: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n" +Feb 25 10:31:41.384: INFO: stdout: "\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p" +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p +Feb 25 10:31:41.384: INFO: Cleaning up the exec pod +STEP: deleting ReplicationController affinity-clusterip-transition in namespace services-8674, will wait for the garbage collector to delete the pods 02/25/23 10:31:41.447 +Feb 25 10:31:41.545: INFO: Deleting ReplicationController affinity-clusterip-transition took: 32.876024ms +Feb 25 10:31:41.646: INFO: Terminating ReplicationController affinity-clusterip-transition pods took: 100.786906ms +[AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 16:30:38.973: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "disruption-9548" for this suite. 12/14/22 16:30:38.981 -{"msg":"PASSED [sig-apps] DisruptionController should block an eviction until the PDB is updated to allow it [Conformance]","completed":154,"skipped":2778,"failed":0} +Feb 25 10:31:44.088: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-8674" for this suite. 02/25/23 10:31:44.103 +[AfterEach] [sig-network] Services + test/e2e/network/service.go:762 +{"msg":"PASSED [sig-network] Services should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance]","completed":148,"skipped":2671,"failed":0} ------------------------------ -• [SLOW TEST] [8.377 seconds] -[sig-apps] DisruptionController -test/e2e/apps/framework.go:23 - should block an eviction until the PDB is updated to allow it [Conformance] - test/e2e/apps/disruption.go:346 +• [SLOW TEST] [13.561 seconds] +[sig-network] Services +test/e2e/network/common/framework.go:23 + should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance] + test/e2e/network/service.go:2189 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] DisruptionController + [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:30:30.672 - Dec 14 16:30:30.672: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename disruption 12/14/22 16:30:30.674 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:30.698 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:30.702 - [BeforeEach] [sig-apps] DisruptionController - test/e2e/apps/disruption.go:71 - [It] should block an eviction until the PDB is updated to allow it [Conformance] - test/e2e/apps/disruption.go:346 - STEP: Creating a pdb that targets all three pods in a test replica set 12/14/22 16:30:30.705 - STEP: Waiting for the pdb to be processed 12/14/22 16:30:30.712 - STEP: First trying to evict a pod which shouldn't be evictable 12/14/22 16:30:32.733 - STEP: Waiting for all pods to be running 12/14/22 16:30:32.733 - Dec 14 16:30:32.740: INFO: pods: 0 < 3 - STEP: locating a running pod 12/14/22 16:30:34.749 - STEP: Updating the pdb to allow a pod to be evicted 12/14/22 16:30:34.768 - STEP: Waiting for the pdb to be processed 12/14/22 16:30:34.781 - STEP: Trying to evict the same pod we tried earlier which should now be evictable 12/14/22 16:30:36.797 - STEP: Waiting for all pods to be running 12/14/22 16:30:36.797 - STEP: Waiting for the pdb to observed all healthy pods 12/14/22 16:30:36.806 - STEP: Patching the pdb to disallow a pod to be evicted 12/14/22 16:30:36.838 - STEP: Waiting for the pdb to be processed 12/14/22 16:30:36.875 - STEP: Waiting for all pods to be running 12/14/22 16:30:38.914 - STEP: locating a running pod 12/14/22 16:30:38.922 - STEP: Deleting the pdb to allow a pod to be evicted 12/14/22 16:30:38.939 - STEP: Waiting for the pdb to be deleted 12/14/22 16:30:38.948 - STEP: Trying to evict the same pod we tried earlier which should now be evictable 12/14/22 16:30:38.952 - STEP: Waiting for all pods to be running 12/14/22 16:30:38.952 - [AfterEach] [sig-apps] DisruptionController + STEP: Creating a kubernetes client 02/25/23 10:31:30.578 + Feb 25 10:31:30.579: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 10:31:30.593 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:31:30.636 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:31:30.64 + [BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 + [It] should be able to switch session affinity for service with type clusterIP [LinuxOnly] [Conformance] + test/e2e/network/service.go:2189 + STEP: creating service in namespace services-8674 02/25/23 10:31:30.646 + STEP: creating service affinity-clusterip-transition in namespace services-8674 02/25/23 10:31:30.647 + STEP: creating replication controller affinity-clusterip-transition in namespace services-8674 02/25/23 10:31:30.685 + I0225 10:31:30.717266 13 runners.go:193] Created replication controller with name: affinity-clusterip-transition, namespace: services-8674, replica count: 3 + I0225 10:31:33.768298 13 runners.go:193] affinity-clusterip-transition Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + Feb 25 10:31:33.785: INFO: Creating new exec pod + Feb 25 10:31:33.812: INFO: Waiting up to 5m0s for pod "execpod-affinityfp4mz" in namespace "services-8674" to be "running" + Feb 25 10:31:33.833: INFO: Pod "execpod-affinityfp4mz": Phase="Pending", Reason="", readiness=false. Elapsed: 21.292329ms + Feb 25 10:31:35.841: INFO: Pod "execpod-affinityfp4mz": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02892199s + Feb 25 10:31:37.844: INFO: Pod "execpod-affinityfp4mz": Phase="Running", Reason="", readiness=true. Elapsed: 4.031626635s + Feb 25 10:31:37.844: INFO: Pod "execpod-affinityfp4mz" satisfied condition "running" + Feb 25 10:31:38.845: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-8674 exec execpod-affinityfp4mz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-clusterip-transition 80' + Feb 25 10:31:39.289: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-clusterip-transition 80\nConnection to affinity-clusterip-transition 80 port [tcp/http] succeeded!\n" + Feb 25 10:31:39.289: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 10:31:39.289: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-8674 exec execpod-affinityfp4mz -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.12.194 80' + Feb 25 10:31:39.555: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.12.194 80\nConnection to 10.233.12.194 80 port [tcp/http] succeeded!\n" + Feb 25 10:31:39.555: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 10:31:39.582: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-8674 exec execpod-affinityfp4mz -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.12.194:80/ ; done' + Feb 25 10:31:40.521: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n" + Feb 25 10:31:40.522: INFO: stdout: "\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-gpgkh\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-gpgkh\naffinity-clusterip-transition-gpgkh\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-q7c8w\naffinity-clusterip-transition-q7c8w\naffinity-clusterip-transition-gpgkh\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-q7c8w\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p" + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-gpgkh + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-gpgkh + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-gpgkh + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-q7c8w + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-q7c8w + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-gpgkh + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-q7c8w + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:40.522: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:40.595: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-8674 exec execpod-affinityfp4mz -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.12.194:80/ ; done' + Feb 25 10:31:41.384: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.12.194:80/\n" + Feb 25 10:31:41.384: INFO: stdout: "\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p\naffinity-clusterip-transition-djj8p" + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Received response from host: affinity-clusterip-transition-djj8p + Feb 25 10:31:41.384: INFO: Cleaning up the exec pod + STEP: deleting ReplicationController affinity-clusterip-transition in namespace services-8674, will wait for the garbage collector to delete the pods 02/25/23 10:31:41.447 + Feb 25 10:31:41.545: INFO: Deleting ReplicationController affinity-clusterip-transition took: 32.876024ms + Feb 25 10:31:41.646: INFO: Terminating ReplicationController affinity-clusterip-transition pods took: 100.786906ms + [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 16:30:38.973: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "disruption-9548" for this suite. 12/14/22 16:30:38.981 + Feb 25 10:31:44.088: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-8674" for this suite. 02/25/23 10:31:44.103 + [AfterEach] [sig-network] Services + test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -[sig-cli] Kubectl client Update Demo - should create and stop a replication controller [Conformance] - test/e2e/kubectl/kubectl.go:337 -[BeforeEach] [sig-cli] Kubectl client +SSSSS +------------------------------ +[sig-api-machinery] Garbage collector + should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] + test/e2e/apimachinery/garbage_collector.go:550 +[BeforeEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:30:39.051 -Dec 14 16:30:39.052: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 16:30:39.054 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:39.171 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:39.175 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 -[BeforeEach] Update Demo - test/e2e/kubectl/kubectl.go:324 -[It] should create and stop a replication controller [Conformance] - test/e2e/kubectl/kubectl.go:337 -STEP: creating a replication controller 12/14/22 16:30:39.179 -Dec 14 16:30:39.180: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 create -f -' -Dec 14 16:30:40.444: INFO: stderr: "" -Dec 14 16:30:40.444: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" -STEP: waiting for all containers in name=update-demo pods to come up. 12/14/22 16:30:40.444 -Dec 14 16:30:40.446: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' -Dec 14 16:30:40.667: INFO: stderr: "" -Dec 14 16:30:40.667: INFO: stdout: "update-demo-nautilus-bf77b update-demo-nautilus-fwbxf " -Dec 14 16:30:40.667: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods update-demo-nautilus-bf77b -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' -Dec 14 16:30:40.823: INFO: stderr: "" -Dec 14 16:30:40.823: INFO: stdout: "" -Dec 14 16:30:40.823: INFO: update-demo-nautilus-bf77b is created but not running -Dec 14 16:30:45.825: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' -Dec 14 16:30:46.016: INFO: stderr: "" -Dec 14 16:30:46.017: INFO: stdout: "update-demo-nautilus-bf77b update-demo-nautilus-fwbxf " -Dec 14 16:30:46.017: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods update-demo-nautilus-bf77b -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' -Dec 14 16:30:46.178: INFO: stderr: "" -Dec 14 16:30:46.178: INFO: stdout: "" -Dec 14 16:30:46.178: INFO: update-demo-nautilus-bf77b is created but not running -Dec 14 16:30:51.179: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' -Dec 14 16:30:51.303: INFO: stderr: "" -Dec 14 16:30:51.303: INFO: stdout: "update-demo-nautilus-bf77b update-demo-nautilus-fwbxf " -Dec 14 16:30:51.303: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods update-demo-nautilus-bf77b -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' -Dec 14 16:30:51.403: INFO: stderr: "" -Dec 14 16:30:51.403: INFO: stdout: "true" -Dec 14 16:30:51.403: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods update-demo-nautilus-bf77b -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' -Dec 14 16:30:51.513: INFO: stderr: "" -Dec 14 16:30:51.513: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" -Dec 14 16:30:51.513: INFO: validating pod update-demo-nautilus-bf77b -Dec 14 16:30:51.529: INFO: got data: { - "image": "nautilus.jpg" -} - -Dec 14 16:30:51.529: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -Dec 14 16:30:51.529: INFO: update-demo-nautilus-bf77b is verified up and running -Dec 14 16:30:51.530: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods update-demo-nautilus-fwbxf -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' -Dec 14 16:30:51.651: INFO: stderr: "" -Dec 14 16:30:51.651: INFO: stdout: "true" -Dec 14 16:30:51.651: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods update-demo-nautilus-fwbxf -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' -Dec 14 16:30:51.793: INFO: stderr: "" -Dec 14 16:30:51.793: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" -Dec 14 16:30:51.793: INFO: validating pod update-demo-nautilus-fwbxf -Dec 14 16:30:51.817: INFO: got data: { - "image": "nautilus.jpg" -} +STEP: Creating a kubernetes client 02/25/23 10:31:44.145 +Feb 25 10:31:44.145: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename gc 02/25/23 10:31:44.15 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:31:44.188 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:31:44.2 +[It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] + test/e2e/apimachinery/garbage_collector.go:550 +STEP: create the deployment 02/25/23 10:31:44.208 +STEP: Wait for the Deployment to create new ReplicaSet 02/25/23 10:31:44.224 +STEP: delete the deployment 02/25/23 10:31:44.393 +STEP: wait for deployment deletion to see if the garbage collector mistakenly deletes the rs 02/25/23 10:31:44.439 +STEP: Gathering metrics 02/25/23 10:31:45.089 +Feb 25 10:31:45.318: INFO: Waiting up to 5m0s for pod "kube-controller-manager-diek3dua9gay-2" in namespace "kube-system" to be "running and ready" +Feb 25 10:31:45.451: INFO: Pod "kube-controller-manager-diek3dua9gay-2": Phase="Running", Reason="", readiness=true. Elapsed: 133.019116ms +Feb 25 10:31:45.451: INFO: The phase of Pod kube-controller-manager-diek3dua9gay-2 is Running (Ready = true) +Feb 25 10:31:45.451: INFO: Pod "kube-controller-manager-diek3dua9gay-2" satisfied condition "running and ready" +Feb 25 10:31:45.762: INFO: For apiserver_request_total: +For apiserver_request_latency_seconds: +For apiserver_init_events_total: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: -Dec 14 16:30:51.818: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -Dec 14 16:30:51.818: INFO: update-demo-nautilus-fwbxf is verified up and running -STEP: using delete to clean up resources 12/14/22 16:30:51.818 -Dec 14 16:30:51.819: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 delete --grace-period=0 --force -f -' -Dec 14 16:30:52.097: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Dec 14 16:30:52.097: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" -Dec 14 16:30:52.097: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get rc,svc -l name=update-demo --no-headers' -Dec 14 16:30:52.461: INFO: stderr: "No resources found in kubectl-8531 namespace.\n" -Dec 14 16:30:52.461: INFO: stdout: "" -Dec 14 16:30:52.461: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods -l name=update-demo -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' -Dec 14 16:30:52.667: INFO: stderr: "" -Dec 14 16:30:52.667: INFO: stdout: "" -[AfterEach] [sig-cli] Kubectl client +[AfterEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:187 -Dec 14 16:30:52.667: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-8531" for this suite. 12/14/22 16:30:52.676 -{"msg":"PASSED [sig-cli] Kubectl client Update Demo should create and stop a replication controller [Conformance]","completed":155,"skipped":2778,"failed":0} +Feb 25 10:31:45.769: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "gc-5103" for this suite. 02/25/23 10:31:45.78 +{"msg":"PASSED [sig-api-machinery] Garbage collector should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance]","completed":149,"skipped":2676,"failed":0} ------------------------------ -• [SLOW TEST] [13.634 seconds] -[sig-cli] Kubectl client -test/e2e/kubectl/framework.go:23 - Update Demo - test/e2e/kubectl/kubectl.go:322 - should create and stop a replication controller [Conformance] - test/e2e/kubectl/kubectl.go:337 +• [1.661 seconds] +[sig-api-machinery] Garbage collector +test/e2e/apimachinery/framework.go:23 + should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] + test/e2e/apimachinery/garbage_collector.go:550 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-cli] Kubectl client + [BeforeEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:30:39.051 - Dec 14 16:30:39.052: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 16:30:39.054 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:39.171 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:39.175 - [BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 - [BeforeEach] Update Demo - test/e2e/kubectl/kubectl.go:324 - [It] should create and stop a replication controller [Conformance] - test/e2e/kubectl/kubectl.go:337 - STEP: creating a replication controller 12/14/22 16:30:39.179 - Dec 14 16:30:39.180: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 create -f -' - Dec 14 16:30:40.444: INFO: stderr: "" - Dec 14 16:30:40.444: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" - STEP: waiting for all containers in name=update-demo pods to come up. 12/14/22 16:30:40.444 - Dec 14 16:30:40.446: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' - Dec 14 16:30:40.667: INFO: stderr: "" - Dec 14 16:30:40.667: INFO: stdout: "update-demo-nautilus-bf77b update-demo-nautilus-fwbxf " - Dec 14 16:30:40.667: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods update-demo-nautilus-bf77b -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' - Dec 14 16:30:40.823: INFO: stderr: "" - Dec 14 16:30:40.823: INFO: stdout: "" - Dec 14 16:30:40.823: INFO: update-demo-nautilus-bf77b is created but not running - Dec 14 16:30:45.825: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' - Dec 14 16:30:46.016: INFO: stderr: "" - Dec 14 16:30:46.017: INFO: stdout: "update-demo-nautilus-bf77b update-demo-nautilus-fwbxf " - Dec 14 16:30:46.017: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods update-demo-nautilus-bf77b -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' - Dec 14 16:30:46.178: INFO: stderr: "" - Dec 14 16:30:46.178: INFO: stdout: "" - Dec 14 16:30:46.178: INFO: update-demo-nautilus-bf77b is created but not running - Dec 14 16:30:51.179: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' - Dec 14 16:30:51.303: INFO: stderr: "" - Dec 14 16:30:51.303: INFO: stdout: "update-demo-nautilus-bf77b update-demo-nautilus-fwbxf " - Dec 14 16:30:51.303: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods update-demo-nautilus-bf77b -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' - Dec 14 16:30:51.403: INFO: stderr: "" - Dec 14 16:30:51.403: INFO: stdout: "true" - Dec 14 16:30:51.403: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods update-demo-nautilus-bf77b -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' - Dec 14 16:30:51.513: INFO: stderr: "" - Dec 14 16:30:51.513: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" - Dec 14 16:30:51.513: INFO: validating pod update-demo-nautilus-bf77b - Dec 14 16:30:51.529: INFO: got data: { - "image": "nautilus.jpg" - } - - Dec 14 16:30:51.529: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . - Dec 14 16:30:51.529: INFO: update-demo-nautilus-bf77b is verified up and running - Dec 14 16:30:51.530: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods update-demo-nautilus-fwbxf -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' - Dec 14 16:30:51.651: INFO: stderr: "" - Dec 14 16:30:51.651: INFO: stdout: "true" - Dec 14 16:30:51.651: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods update-demo-nautilus-fwbxf -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' - Dec 14 16:30:51.793: INFO: stderr: "" - Dec 14 16:30:51.793: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" - Dec 14 16:30:51.793: INFO: validating pod update-demo-nautilus-fwbxf - Dec 14 16:30:51.817: INFO: got data: { - "image": "nautilus.jpg" - } + STEP: Creating a kubernetes client 02/25/23 10:31:44.145 + Feb 25 10:31:44.145: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename gc 02/25/23 10:31:44.15 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:31:44.188 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:31:44.2 + [It] should orphan RS created by deployment when deleteOptions.PropagationPolicy is Orphan [Conformance] + test/e2e/apimachinery/garbage_collector.go:550 + STEP: create the deployment 02/25/23 10:31:44.208 + STEP: Wait for the Deployment to create new ReplicaSet 02/25/23 10:31:44.224 + STEP: delete the deployment 02/25/23 10:31:44.393 + STEP: wait for deployment deletion to see if the garbage collector mistakenly deletes the rs 02/25/23 10:31:44.439 + STEP: Gathering metrics 02/25/23 10:31:45.089 + Feb 25 10:31:45.318: INFO: Waiting up to 5m0s for pod "kube-controller-manager-diek3dua9gay-2" in namespace "kube-system" to be "running and ready" + Feb 25 10:31:45.451: INFO: Pod "kube-controller-manager-diek3dua9gay-2": Phase="Running", Reason="", readiness=true. Elapsed: 133.019116ms + Feb 25 10:31:45.451: INFO: The phase of Pod kube-controller-manager-diek3dua9gay-2 is Running (Ready = true) + Feb 25 10:31:45.451: INFO: Pod "kube-controller-manager-diek3dua9gay-2" satisfied condition "running and ready" + Feb 25 10:31:45.762: INFO: For apiserver_request_total: + For apiserver_request_latency_seconds: + For apiserver_init_events_total: + For garbage_collector_attempt_to_delete_queue_latency: + For garbage_collector_attempt_to_delete_work_duration: + For garbage_collector_attempt_to_orphan_queue_latency: + For garbage_collector_attempt_to_orphan_work_duration: + For garbage_collector_dirty_processing_latency_microseconds: + For garbage_collector_event_processing_latency_microseconds: + For garbage_collector_graph_changes_queue_latency: + For garbage_collector_graph_changes_work_duration: + For garbage_collector_orphan_processing_latency_microseconds: + For namespace_queue_latency: + For namespace_queue_latency_sum: + For namespace_queue_latency_count: + For namespace_retries: + For namespace_work_duration: + For namespace_work_duration_sum: + For namespace_work_duration_count: + For function_duration_seconds: + For errors_total: + For evicted_pods_total: - Dec 14 16:30:51.818: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . - Dec 14 16:30:51.818: INFO: update-demo-nautilus-fwbxf is verified up and running - STEP: using delete to clean up resources 12/14/22 16:30:51.818 - Dec 14 16:30:51.819: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 delete --grace-period=0 --force -f -' - Dec 14 16:30:52.097: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" - Dec 14 16:30:52.097: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" - Dec 14 16:30:52.097: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get rc,svc -l name=update-demo --no-headers' - Dec 14 16:30:52.461: INFO: stderr: "No resources found in kubectl-8531 namespace.\n" - Dec 14 16:30:52.461: INFO: stdout: "" - Dec 14 16:30:52.461: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-8531 get pods -l name=update-demo -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' - Dec 14 16:30:52.667: INFO: stderr: "" - Dec 14 16:30:52.667: INFO: stdout: "" - [AfterEach] [sig-cli] Kubectl client + [AfterEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:187 - Dec 14 16:30:52.667: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-8531" for this suite. 12/14/22 16:30:52.676 + Feb 25 10:31:45.769: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "gc-5103" for this suite. 02/25/23 10:31:45.78 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSS ------------------------------ -[sig-apps] ReplicaSet - should list and delete a collection of ReplicaSets [Conformance] - test/e2e/apps/replica_set.go:165 -[BeforeEach] [sig-apps] ReplicaSet +[sig-storage] Projected secret + should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:55 +[BeforeEach] [sig-storage] Projected secret test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:30:52.691 -Dec 14 16:30:52.691: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename replicaset 12/14/22 16:30:52.696 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:52.725 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:52.729 -[It] should list and delete a collection of ReplicaSets [Conformance] - test/e2e/apps/replica_set.go:165 -STEP: Create a ReplicaSet 12/14/22 16:30:52.732 -STEP: Verify that the required pods have come up 12/14/22 16:30:52.739 -Dec 14 16:30:52.744: INFO: Pod name sample-pod: Found 0 pods out of 3 -Dec 14 16:30:57.792: INFO: Pod name sample-pod: Found 3 pods out of 3 -STEP: ensuring each pod is running 12/14/22 16:30:57.792 -Dec 14 16:30:57.818: INFO: Replica Status: {Replicas:3 FullyLabeledReplicas:3 ReadyReplicas:3 AvailableReplicas:3 ObservedGeneration:1 Conditions:[]} -STEP: Listing all ReplicaSets 12/14/22 16:30:57.818 -STEP: DeleteCollection of the ReplicaSets 12/14/22 16:30:57.841 -STEP: After DeleteCollection verify that ReplicaSets have been deleted 12/14/22 16:30:57.918 -[AfterEach] [sig-apps] ReplicaSet +STEP: Creating a kubernetes client 02/25/23 10:31:45.811 +Feb 25 10:31:45.811: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 10:31:45.825 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:31:45.87 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:31:45.877 +[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:55 +STEP: Creating projection with secret that has name projected-secret-test-9c565b11-e54d-4b15-a0cb-823cb3ee3370 02/25/23 10:31:45.882 +STEP: Creating a pod to test consume secrets 02/25/23 10:31:45.893 +Feb 25 10:31:45.921: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7" in namespace "projected-2082" to be "Succeeded or Failed" +Feb 25 10:31:45.937: INFO: Pod "pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7": Phase="Pending", Reason="", readiness=false. Elapsed: 16.007304ms +Feb 25 10:31:47.948: INFO: Pod "pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02719677s +Feb 25 10:31:49.957: INFO: Pod "pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7": Phase="Pending", Reason="", readiness=false. Elapsed: 4.035951296s +Feb 25 10:31:51.943: INFO: Pod "pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.022605187s +STEP: Saw pod success 02/25/23 10:31:51.943 +Feb 25 10:31:51.944: INFO: Pod "pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7" satisfied condition "Succeeded or Failed" +Feb 25 10:31:51.951: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7 container projected-secret-volume-test: +STEP: delete the pod 02/25/23 10:31:52.001 +Feb 25 10:31:52.034: INFO: Waiting for pod pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7 to disappear +Feb 25 10:31:52.045: INFO: Pod pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7 no longer exists +[AfterEach] [sig-storage] Projected secret test/e2e/framework/framework.go:187 -Dec 14 16:30:57.937: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "replicaset-6010" for this suite. 12/14/22 16:30:57.987 -{"msg":"PASSED [sig-apps] ReplicaSet should list and delete a collection of ReplicaSets [Conformance]","completed":156,"skipped":2831,"failed":0} +Feb 25 10:31:52.045: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-2082" for this suite. 02/25/23 10:31:52.067 +{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","completed":150,"skipped":2691,"failed":0} ------------------------------ -• [SLOW TEST] [5.329 seconds] -[sig-apps] ReplicaSet -test/e2e/apps/framework.go:23 - should list and delete a collection of ReplicaSets [Conformance] - test/e2e/apps/replica_set.go:165 +• [SLOW TEST] [6.284 seconds] +[sig-storage] Projected secret +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:55 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] ReplicaSet + [BeforeEach] [sig-storage] Projected secret test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:30:52.691 - Dec 14 16:30:52.691: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename replicaset 12/14/22 16:30:52.696 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:52.725 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:52.729 - [It] should list and delete a collection of ReplicaSets [Conformance] - test/e2e/apps/replica_set.go:165 - STEP: Create a ReplicaSet 12/14/22 16:30:52.732 - STEP: Verify that the required pods have come up 12/14/22 16:30:52.739 - Dec 14 16:30:52.744: INFO: Pod name sample-pod: Found 0 pods out of 3 - Dec 14 16:30:57.792: INFO: Pod name sample-pod: Found 3 pods out of 3 - STEP: ensuring each pod is running 12/14/22 16:30:57.792 - Dec 14 16:30:57.818: INFO: Replica Status: {Replicas:3 FullyLabeledReplicas:3 ReadyReplicas:3 AvailableReplicas:3 ObservedGeneration:1 Conditions:[]} - STEP: Listing all ReplicaSets 12/14/22 16:30:57.818 - STEP: DeleteCollection of the ReplicaSets 12/14/22 16:30:57.841 - STEP: After DeleteCollection verify that ReplicaSets have been deleted 12/14/22 16:30:57.918 - [AfterEach] [sig-apps] ReplicaSet + STEP: Creating a kubernetes client 02/25/23 10:31:45.811 + Feb 25 10:31:45.811: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 10:31:45.825 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:31:45.87 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:31:45.877 + [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:55 + STEP: Creating projection with secret that has name projected-secret-test-9c565b11-e54d-4b15-a0cb-823cb3ee3370 02/25/23 10:31:45.882 + STEP: Creating a pod to test consume secrets 02/25/23 10:31:45.893 + Feb 25 10:31:45.921: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7" in namespace "projected-2082" to be "Succeeded or Failed" + Feb 25 10:31:45.937: INFO: Pod "pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7": Phase="Pending", Reason="", readiness=false. Elapsed: 16.007304ms + Feb 25 10:31:47.948: INFO: Pod "pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02719677s + Feb 25 10:31:49.957: INFO: Pod "pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7": Phase="Pending", Reason="", readiness=false. Elapsed: 4.035951296s + Feb 25 10:31:51.943: INFO: Pod "pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.022605187s + STEP: Saw pod success 02/25/23 10:31:51.943 + Feb 25 10:31:51.944: INFO: Pod "pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7" satisfied condition "Succeeded or Failed" + Feb 25 10:31:51.951: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7 container projected-secret-volume-test: + STEP: delete the pod 02/25/23 10:31:52.001 + Feb 25 10:31:52.034: INFO: Waiting for pod pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7 to disappear + Feb 25 10:31:52.045: INFO: Pod pod-projected-secrets-fe779c69-8551-4625-9fcd-2371675a9ef7 no longer exists + [AfterEach] [sig-storage] Projected secret test/e2e/framework/framework.go:187 - Dec 14 16:30:57.937: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "replicaset-6010" for this suite. 12/14/22 16:30:57.987 + Feb 25 10:31:52.045: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-2082" for this suite. 02/25/23 10:31:52.067 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSS +SS ------------------------------ -[sig-storage] Downward API volume - should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:260 -[BeforeEach] [sig-storage] Downward API volume +[sig-node] Probing container + should have monotonically increasing restart count [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:195 +[BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:30:58.032 -Dec 14 16:30:58.032: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 16:30:58.041 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:58.072 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:58.077 -[BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 -[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:260 -STEP: Creating a pod to test downward API volume plugin 12/14/22 16:30:58.083 -Dec 14 16:30:58.097: INFO: Waiting up to 5m0s for pod "downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594" in namespace "downward-api-9476" to be "Succeeded or Failed" -Dec 14 16:30:58.103: INFO: Pod "downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594": Phase="Pending", Reason="", readiness=false. Elapsed: 5.711372ms -Dec 14 16:31:00.108: INFO: Pod "downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010267463s -Dec 14 16:31:02.108: INFO: Pod "downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010943731s -STEP: Saw pod success 12/14/22 16:31:02.108 -Dec 14 16:31:02.109: INFO: Pod "downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594" satisfied condition "Succeeded or Failed" -Dec 14 16:31:02.114: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594 container client-container: -STEP: delete the pod 12/14/22 16:31:02.138 -Dec 14 16:31:02.154: INFO: Waiting for pod downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594 to disappear -Dec 14 16:31:02.158: INFO: Pod downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594 no longer exists -[AfterEach] [sig-storage] Downward API volume +STEP: Creating a kubernetes client 02/25/23 10:31:52.096 +Feb 25 10:31:52.096: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-probe 02/25/23 10:31:52.098 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:31:52.133 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:31:52.137 +[BeforeEach] [sig-node] Probing container + test/e2e/common/node/container_probe.go:59 +[It] should have monotonically increasing restart count [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:195 +STEP: Creating pod liveness-67171baf-e478-4d2c-a7e7-555281caaa04 in namespace container-probe-9096 02/25/23 10:31:52.144 +Feb 25 10:31:52.175: INFO: Waiting up to 5m0s for pod "liveness-67171baf-e478-4d2c-a7e7-555281caaa04" in namespace "container-probe-9096" to be "not pending" +Feb 25 10:31:52.183: INFO: Pod "liveness-67171baf-e478-4d2c-a7e7-555281caaa04": Phase="Pending", Reason="", readiness=false. Elapsed: 7.630632ms +Feb 25 10:31:54.190: INFO: Pod "liveness-67171baf-e478-4d2c-a7e7-555281caaa04": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01484368s +Feb 25 10:31:56.191: INFO: Pod "liveness-67171baf-e478-4d2c-a7e7-555281caaa04": Phase="Running", Reason="", readiness=true. Elapsed: 4.016105772s +Feb 25 10:31:56.192: INFO: Pod "liveness-67171baf-e478-4d2c-a7e7-555281caaa04" satisfied condition "not pending" +Feb 25 10:31:56.192: INFO: Started pod liveness-67171baf-e478-4d2c-a7e7-555281caaa04 in namespace container-probe-9096 +STEP: checking the pod's current state and verifying that restartCount is present 02/25/23 10:31:56.192 +Feb 25 10:31:56.199: INFO: Initial restart count of pod liveness-67171baf-e478-4d2c-a7e7-555281caaa04 is 0 +Feb 25 10:32:14.301: INFO: Restart count of pod container-probe-9096/liveness-67171baf-e478-4d2c-a7e7-555281caaa04 is now 1 (18.101915775s elapsed) +Feb 25 10:32:34.381: INFO: Restart count of pod container-probe-9096/liveness-67171baf-e478-4d2c-a7e7-555281caaa04 is now 2 (38.182564179s elapsed) +Feb 25 10:32:54.481: INFO: Restart count of pod container-probe-9096/liveness-67171baf-e478-4d2c-a7e7-555281caaa04 is now 3 (58.282233222s elapsed) +Feb 25 10:33:14.582: INFO: Restart count of pod container-probe-9096/liveness-67171baf-e478-4d2c-a7e7-555281caaa04 is now 4 (1m18.38318031s elapsed) +Feb 25 10:34:16.879: INFO: Restart count of pod container-probe-9096/liveness-67171baf-e478-4d2c-a7e7-555281caaa04 is now 5 (2m20.679971337s elapsed) +STEP: deleting the pod 02/25/23 10:34:16.879 +[AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 -Dec 14 16:31:02.158: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-9476" for this suite. 12/14/22 16:31:02.164 -{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","completed":157,"skipped":2840,"failed":0} +Feb 25 10:34:16.987: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-9096" for this suite. 02/25/23 10:34:17.041 +{"msg":"PASSED [sig-node] Probing container should have monotonically increasing restart count [NodeConformance] [Conformance]","completed":151,"skipped":2693,"failed":0} ------------------------------ -• [4.141 seconds] -[sig-storage] Downward API volume -test/e2e/common/storage/framework.go:23 - should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:260 +• [SLOW TEST] [144.975 seconds] +[sig-node] Probing container +test/e2e/common/node/framework.go:23 + should have monotonically increasing restart count [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:195 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Downward API volume + [BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:30:58.032 - Dec 14 16:30:58.032: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 16:30:58.041 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:30:58.072 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:30:58.077 - [BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 - [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:260 - STEP: Creating a pod to test downward API volume plugin 12/14/22 16:30:58.083 - Dec 14 16:30:58.097: INFO: Waiting up to 5m0s for pod "downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594" in namespace "downward-api-9476" to be "Succeeded or Failed" - Dec 14 16:30:58.103: INFO: Pod "downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594": Phase="Pending", Reason="", readiness=false. Elapsed: 5.711372ms - Dec 14 16:31:00.108: INFO: Pod "downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010267463s - Dec 14 16:31:02.108: INFO: Pod "downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010943731s - STEP: Saw pod success 12/14/22 16:31:02.108 - Dec 14 16:31:02.109: INFO: Pod "downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594" satisfied condition "Succeeded or Failed" - Dec 14 16:31:02.114: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594 container client-container: - STEP: delete the pod 12/14/22 16:31:02.138 - Dec 14 16:31:02.154: INFO: Waiting for pod downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594 to disappear - Dec 14 16:31:02.158: INFO: Pod downwardapi-volume-97545dba-ffb6-4d1a-8c4b-0af318d73594 no longer exists - [AfterEach] [sig-storage] Downward API volume + STEP: Creating a kubernetes client 02/25/23 10:31:52.096 + Feb 25 10:31:52.096: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-probe 02/25/23 10:31:52.098 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:31:52.133 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:31:52.137 + [BeforeEach] [sig-node] Probing container + test/e2e/common/node/container_probe.go:59 + [It] should have monotonically increasing restart count [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:195 + STEP: Creating pod liveness-67171baf-e478-4d2c-a7e7-555281caaa04 in namespace container-probe-9096 02/25/23 10:31:52.144 + Feb 25 10:31:52.175: INFO: Waiting up to 5m0s for pod "liveness-67171baf-e478-4d2c-a7e7-555281caaa04" in namespace "container-probe-9096" to be "not pending" + Feb 25 10:31:52.183: INFO: Pod "liveness-67171baf-e478-4d2c-a7e7-555281caaa04": Phase="Pending", Reason="", readiness=false. Elapsed: 7.630632ms + Feb 25 10:31:54.190: INFO: Pod "liveness-67171baf-e478-4d2c-a7e7-555281caaa04": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01484368s + Feb 25 10:31:56.191: INFO: Pod "liveness-67171baf-e478-4d2c-a7e7-555281caaa04": Phase="Running", Reason="", readiness=true. Elapsed: 4.016105772s + Feb 25 10:31:56.192: INFO: Pod "liveness-67171baf-e478-4d2c-a7e7-555281caaa04" satisfied condition "not pending" + Feb 25 10:31:56.192: INFO: Started pod liveness-67171baf-e478-4d2c-a7e7-555281caaa04 in namespace container-probe-9096 + STEP: checking the pod's current state and verifying that restartCount is present 02/25/23 10:31:56.192 + Feb 25 10:31:56.199: INFO: Initial restart count of pod liveness-67171baf-e478-4d2c-a7e7-555281caaa04 is 0 + Feb 25 10:32:14.301: INFO: Restart count of pod container-probe-9096/liveness-67171baf-e478-4d2c-a7e7-555281caaa04 is now 1 (18.101915775s elapsed) + Feb 25 10:32:34.381: INFO: Restart count of pod container-probe-9096/liveness-67171baf-e478-4d2c-a7e7-555281caaa04 is now 2 (38.182564179s elapsed) + Feb 25 10:32:54.481: INFO: Restart count of pod container-probe-9096/liveness-67171baf-e478-4d2c-a7e7-555281caaa04 is now 3 (58.282233222s elapsed) + Feb 25 10:33:14.582: INFO: Restart count of pod container-probe-9096/liveness-67171baf-e478-4d2c-a7e7-555281caaa04 is now 4 (1m18.38318031s elapsed) + Feb 25 10:34:16.879: INFO: Restart count of pod container-probe-9096/liveness-67171baf-e478-4d2c-a7e7-555281caaa04 is now 5 (2m20.679971337s elapsed) + STEP: deleting the pod 02/25/23 10:34:16.879 + [AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 - Dec 14 16:31:02.158: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-9476" for this suite. 12/14/22 16:31:02.164 + Feb 25 10:34:16.987: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-probe-9096" for this suite. 02/25/23 10:34:17.041 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSS +SSSSSSS ------------------------------ -[sig-network] Services - should be able to create a functioning NodePort service [Conformance] - test/e2e/network/service.go:1268 -[BeforeEach] [sig-network] Services +[sig-apps] Daemon set [Serial] + should rollback without unnecessary restarts [Conformance] + test/e2e/apps/daemon_set.go:431 +[BeforeEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:31:02.175 -Dec 14 16:31:02.176: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 16:31:02.179 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:31:02.204 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:31:02.209 -[BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 -[It] should be able to create a functioning NodePort service [Conformance] - test/e2e/network/service.go:1268 -STEP: creating service nodeport-test with type=NodePort in namespace services-4258 12/14/22 16:31:02.227 -STEP: creating replication controller nodeport-test in namespace services-4258 12/14/22 16:31:02.265 -I1214 16:31:02.302850 14 runners.go:193] Created replication controller with name: nodeport-test, namespace: services-4258, replica count: 2 -I1214 16:31:05.353912 14 runners.go:193] nodeport-test Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -Dec 14 16:31:05.354: INFO: Creating new exec pod -Dec 14 16:31:05.377: INFO: Waiting up to 5m0s for pod "execpodglz42" in namespace "services-4258" to be "running" -Dec 14 16:31:05.385: INFO: Pod "execpodglz42": Phase="Pending", Reason="", readiness=false. Elapsed: 8.208274ms -Dec 14 16:31:07.391: INFO: Pod "execpodglz42": Phase="Running", Reason="", readiness=true. Elapsed: 2.014092637s -Dec 14 16:31:07.391: INFO: Pod "execpodglz42" satisfied condition "running" -Dec 14 16:31:08.401: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4258 exec execpodglz42 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 nodeport-test 80' -Dec 14 16:31:08.799: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 nodeport-test 80\nConnection to nodeport-test 80 port [tcp/http] succeeded!\n" -Dec 14 16:31:08.799: INFO: stdout: "nodeport-test-p77kn" -Dec 14 16:31:08.800: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4258 exec execpodglz42 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.20.148 80' -Dec 14 16:31:09.004: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.20.148 80\nConnection to 10.233.20.148 80 port [tcp/http] succeeded!\n" -Dec 14 16:31:09.004: INFO: stdout: "nodeport-test-p77kn" -Dec 14 16:31:09.004: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4258 exec execpodglz42 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.244 30338' -Dec 14 16:31:09.244: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.244 30338\nConnection to 192.168.121.244 30338 port [tcp/*] succeeded!\n" -Dec 14 16:31:09.244: INFO: stdout: "nodeport-test-p77kn" -Dec 14 16:31:09.245: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4258 exec execpodglz42 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.249 30338' -Dec 14 16:31:09.456: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.249 30338\nConnection to 192.168.121.249 30338 port [tcp/*] succeeded!\n" -Dec 14 16:31:09.457: INFO: stdout: "nodeport-test-p77kn" -[AfterEach] [sig-network] Services +STEP: Creating a kubernetes client 02/25/23 10:34:17.075 +Feb 25 10:34:17.075: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename daemonsets 02/25/23 10:34:17.081 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:34:17.118 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:34:17.124 +[BeforeEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:145 +[It] should rollback without unnecessary restarts [Conformance] + test/e2e/apps/daemon_set.go:431 +Feb 25 10:34:17.345: INFO: Create a RollingUpdate DaemonSet +Feb 25 10:34:17.368: INFO: Check that daemon pods launch on every node of the cluster +Feb 25 10:34:17.399: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:34:17.399: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 10:34:18.436: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:34:18.436: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 10:34:19.445: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 +Feb 25 10:34:19.446: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 +Feb 25 10:34:20.417: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 +Feb 25 10:34:20.417: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set +Feb 25 10:34:20.417: INFO: Update the DaemonSet to trigger a rollout +Feb 25 10:34:20.440: INFO: Updating DaemonSet daemon-set +Feb 25 10:34:23.490: INFO: Roll back the DaemonSet before rollout is complete +Feb 25 10:34:23.516: INFO: Updating DaemonSet daemon-set +Feb 25 10:34:23.516: INFO: Make sure DaemonSet rollback is complete +Feb 25 10:34:23.529: INFO: Wrong image for pod: daemon-set-46p99. Expected: registry.k8s.io/e2e-test-images/httpd:2.4.38-2, got: foo:non-existent. +Feb 25 10:34:23.529: INFO: Pod daemon-set-46p99 is not available +Feb 25 10:34:30.566: INFO: Pod daemon-set-hbxd5 is not available +[AfterEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:110 +STEP: Deleting DaemonSet "daemon-set" 02/25/23 10:34:30.601 +STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6930, will wait for the garbage collector to delete the pods 02/25/23 10:34:30.601 +Feb 25 10:34:30.677: INFO: Deleting DaemonSet.extensions daemon-set took: 17.359418ms +Feb 25 10:34:30.777: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.346757ms +Feb 25 10:34:32.487: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 10:34:32.487: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set +Feb 25 10:34:32.493: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"24316"},"items":null} + +Feb 25 10:34:32.499: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"24316"},"items":null} + +[AfterEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:187 -Dec 14 16:31:09.457: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-4258" for this suite. 12/14/22 16:31:09.464 -[AfterEach] [sig-network] Services - test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should be able to create a functioning NodePort service [Conformance]","completed":158,"skipped":2852,"failed":0} +Feb 25 10:34:32.531: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "daemonsets-6930" for this suite. 02/25/23 10:34:32.541 +{"msg":"PASSED [sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance]","completed":152,"skipped":2700,"failed":0} ------------------------------ -• [SLOW TEST] [7.299 seconds] -[sig-network] Services -test/e2e/network/common/framework.go:23 - should be able to create a functioning NodePort service [Conformance] - test/e2e/network/service.go:1268 +• [SLOW TEST] [15.481 seconds] +[sig-apps] Daemon set [Serial] +test/e2e/apps/framework.go:23 + should rollback without unnecessary restarts [Conformance] + test/e2e/apps/daemon_set.go:431 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Services + [BeforeEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:31:02.175 - Dec 14 16:31:02.176: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 16:31:02.179 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:31:02.204 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:31:02.209 - [BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 - [It] should be able to create a functioning NodePort service [Conformance] - test/e2e/network/service.go:1268 - STEP: creating service nodeport-test with type=NodePort in namespace services-4258 12/14/22 16:31:02.227 - STEP: creating replication controller nodeport-test in namespace services-4258 12/14/22 16:31:02.265 - I1214 16:31:02.302850 14 runners.go:193] Created replication controller with name: nodeport-test, namespace: services-4258, replica count: 2 - I1214 16:31:05.353912 14 runners.go:193] nodeport-test Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - Dec 14 16:31:05.354: INFO: Creating new exec pod - Dec 14 16:31:05.377: INFO: Waiting up to 5m0s for pod "execpodglz42" in namespace "services-4258" to be "running" - Dec 14 16:31:05.385: INFO: Pod "execpodglz42": Phase="Pending", Reason="", readiness=false. Elapsed: 8.208274ms - Dec 14 16:31:07.391: INFO: Pod "execpodglz42": Phase="Running", Reason="", readiness=true. Elapsed: 2.014092637s - Dec 14 16:31:07.391: INFO: Pod "execpodglz42" satisfied condition "running" - Dec 14 16:31:08.401: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4258 exec execpodglz42 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 nodeport-test 80' - Dec 14 16:31:08.799: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 nodeport-test 80\nConnection to nodeport-test 80 port [tcp/http] succeeded!\n" - Dec 14 16:31:08.799: INFO: stdout: "nodeport-test-p77kn" - Dec 14 16:31:08.800: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4258 exec execpodglz42 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.20.148 80' - Dec 14 16:31:09.004: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.20.148 80\nConnection to 10.233.20.148 80 port [tcp/http] succeeded!\n" - Dec 14 16:31:09.004: INFO: stdout: "nodeport-test-p77kn" - Dec 14 16:31:09.004: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4258 exec execpodglz42 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.244 30338' - Dec 14 16:31:09.244: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.244 30338\nConnection to 192.168.121.244 30338 port [tcp/*] succeeded!\n" - Dec 14 16:31:09.244: INFO: stdout: "nodeport-test-p77kn" - Dec 14 16:31:09.245: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4258 exec execpodglz42 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.249 30338' - Dec 14 16:31:09.456: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.249 30338\nConnection to 192.168.121.249 30338 port [tcp/*] succeeded!\n" - Dec 14 16:31:09.457: INFO: stdout: "nodeport-test-p77kn" - [AfterEach] [sig-network] Services + STEP: Creating a kubernetes client 02/25/23 10:34:17.075 + Feb 25 10:34:17.075: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename daemonsets 02/25/23 10:34:17.081 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:34:17.118 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:34:17.124 + [BeforeEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:145 + [It] should rollback without unnecessary restarts [Conformance] + test/e2e/apps/daemon_set.go:431 + Feb 25 10:34:17.345: INFO: Create a RollingUpdate DaemonSet + Feb 25 10:34:17.368: INFO: Check that daemon pods launch on every node of the cluster + Feb 25 10:34:17.399: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:34:17.399: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 10:34:18.436: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:34:18.436: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 10:34:19.445: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 + Feb 25 10:34:19.446: INFO: Node diek3dua9gay-3 is running 0 daemon pod, expected 1 + Feb 25 10:34:20.417: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 + Feb 25 10:34:20.417: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set + Feb 25 10:34:20.417: INFO: Update the DaemonSet to trigger a rollout + Feb 25 10:34:20.440: INFO: Updating DaemonSet daemon-set + Feb 25 10:34:23.490: INFO: Roll back the DaemonSet before rollout is complete + Feb 25 10:34:23.516: INFO: Updating DaemonSet daemon-set + Feb 25 10:34:23.516: INFO: Make sure DaemonSet rollback is complete + Feb 25 10:34:23.529: INFO: Wrong image for pod: daemon-set-46p99. Expected: registry.k8s.io/e2e-test-images/httpd:2.4.38-2, got: foo:non-existent. + Feb 25 10:34:23.529: INFO: Pod daemon-set-46p99 is not available + Feb 25 10:34:30.566: INFO: Pod daemon-set-hbxd5 is not available + [AfterEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:110 + STEP: Deleting DaemonSet "daemon-set" 02/25/23 10:34:30.601 + STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-6930, will wait for the garbage collector to delete the pods 02/25/23 10:34:30.601 + Feb 25 10:34:30.677: INFO: Deleting DaemonSet.extensions daemon-set took: 17.359418ms + Feb 25 10:34:30.777: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.346757ms + Feb 25 10:34:32.487: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 10:34:32.487: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set + Feb 25 10:34:32.493: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"24316"},"items":null} + + Feb 25 10:34:32.499: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"24316"},"items":null} + + [AfterEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:187 - Dec 14 16:31:09.457: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-4258" for this suite. 12/14/22 16:31:09.464 - [AfterEach] [sig-network] Services - test/e2e/network/service.go:762 + Feb 25 10:34:32.531: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "daemonsets-6930" for this suite. 02/25/23 10:34:32.541 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services - should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance] - test/e2e/network/service.go:2173 + should provide secure master service [Conformance] + test/e2e/network/service.go:781 [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:31:09.479 -Dec 14 16:31:09.479: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 16:31:09.481 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:31:09.505 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:31:09.51 +STEP: Creating a kubernetes client 02/25/23 10:34:32.574 +Feb 25 10:34:32.574: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 10:34:32.577 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:34:32.615 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:34:32.63 [BeforeEach] [sig-network] Services test/e2e/network/service.go:758 -[It] should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance] - test/e2e/network/service.go:2173 -STEP: creating service in namespace services-173 12/14/22 16:31:09.514 -Dec 14 16:31:09.527: INFO: Waiting up to 5m0s for pod "kube-proxy-mode-detector" in namespace "services-173" to be "running and ready" -Dec 14 16:31:09.532: INFO: Pod "kube-proxy-mode-detector": Phase="Pending", Reason="", readiness=false. Elapsed: 4.39727ms -Dec 14 16:31:09.532: INFO: The phase of Pod kube-proxy-mode-detector is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:31:11.541: INFO: Pod "kube-proxy-mode-detector": Phase="Running", Reason="", readiness=true. Elapsed: 2.013554727s -Dec 14 16:31:11.541: INFO: The phase of Pod kube-proxy-mode-detector is Running (Ready = true) -Dec 14 16:31:11.541: INFO: Pod "kube-proxy-mode-detector" satisfied condition "running and ready" -Dec 14 16:31:11.547: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-173 exec kube-proxy-mode-detector -- /bin/sh -x -c curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode' -Dec 14 16:31:11.851: INFO: stderr: "+ curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode\n" -Dec 14 16:31:11.851: INFO: stdout: "iptables" -Dec 14 16:31:11.851: INFO: proxyMode: iptables -Dec 14 16:31:11.872: INFO: Waiting for pod kube-proxy-mode-detector to disappear -Dec 14 16:31:11.876: INFO: Pod kube-proxy-mode-detector no longer exists -STEP: creating service affinity-clusterip-timeout in namespace services-173 12/14/22 16:31:11.876 -STEP: creating replication controller affinity-clusterip-timeout in namespace services-173 12/14/22 16:31:11.894 -I1214 16:31:11.906787 14 runners.go:193] Created replication controller with name: affinity-clusterip-timeout, namespace: services-173, replica count: 3 -I1214 16:31:14.957568 14 runners.go:193] affinity-clusterip-timeout Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -Dec 14 16:31:14.992: INFO: Creating new exec pod -Dec 14 16:31:15.004: INFO: Waiting up to 5m0s for pod "execpod-affinity7vgc4" in namespace "services-173" to be "running" -Dec 14 16:31:15.011: INFO: Pod "execpod-affinity7vgc4": Phase="Pending", Reason="", readiness=false. Elapsed: 7.027481ms -Dec 14 16:31:17.017: INFO: Pod "execpod-affinity7vgc4": Phase="Running", Reason="", readiness=true. Elapsed: 2.012640637s -Dec 14 16:31:17.017: INFO: Pod "execpod-affinity7vgc4" satisfied condition "running" -Dec 14 16:31:18.018: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-173 exec execpod-affinity7vgc4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-clusterip-timeout 80' -Dec 14 16:31:18.320: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-clusterip-timeout 80\nConnection to affinity-clusterip-timeout 80 port [tcp/http] succeeded!\n" -Dec 14 16:31:18.320: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:31:18.321: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-173 exec execpod-affinity7vgc4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.44.168 80' -Dec 14 16:31:18.590: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.44.168 80\nConnection to 10.233.44.168 80 port [tcp/http] succeeded!\n" -Dec 14 16:31:18.590: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:31:18.592: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-173 exec execpod-affinity7vgc4 -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.44.168:80/ ; done' -Dec 14 16:31:19.035: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n" -Dec 14 16:31:19.035: INFO: stdout: "\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4" -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 -Dec 14 16:31:19.035: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-173 exec execpod-affinity7vgc4 -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://10.233.44.168:80/' -Dec 14 16:31:19.243: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n" -Dec 14 16:31:19.243: INFO: stdout: "affinity-clusterip-timeout-qlpt4" -Dec 14 16:31:39.244: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-173 exec execpod-affinity7vgc4 -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://10.233.44.168:80/' -Dec 14 16:31:39.483: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n" -Dec 14 16:31:39.483: INFO: stdout: "affinity-clusterip-timeout-444jf" -Dec 14 16:31:39.483: INFO: Cleaning up the exec pod -STEP: deleting ReplicationController affinity-clusterip-timeout in namespace services-173, will wait for the garbage collector to delete the pods 12/14/22 16:31:39.508 -Dec 14 16:31:39.583: INFO: Deleting ReplicationController affinity-clusterip-timeout took: 13.564809ms -Dec 14 16:31:39.683: INFO: Terminating ReplicationController affinity-clusterip-timeout pods took: 100.496326ms +[It] should provide secure master service [Conformance] + test/e2e/network/service.go:781 [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 16:31:42.044: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-173" for this suite. 12/14/22 16:31:42.054 +Feb 25 10:34:32.653: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-3591" for this suite. 02/25/23 10:34:32.662 [AfterEach] [sig-network] Services test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance]","completed":159,"skipped":2874,"failed":0} +{"msg":"PASSED [sig-network] Services should provide secure master service [Conformance]","completed":153,"skipped":2760,"failed":0} ------------------------------ -• [SLOW TEST] [32.584 seconds] +• [0.100 seconds] [sig-network] Services test/e2e/network/common/framework.go:23 - should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance] - test/e2e/network/service.go:2173 + should provide secure master service [Conformance] + test/e2e/network/service.go:781 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:31:09.479 - Dec 14 16:31:09.479: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 16:31:09.481 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:31:09.505 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:31:09.51 + STEP: Creating a kubernetes client 02/25/23 10:34:32.574 + Feb 25 10:34:32.574: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 10:34:32.577 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:34:32.615 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:34:32.63 [BeforeEach] [sig-network] Services test/e2e/network/service.go:758 - [It] should have session affinity timeout work for service with type clusterIP [LinuxOnly] [Conformance] - test/e2e/network/service.go:2173 - STEP: creating service in namespace services-173 12/14/22 16:31:09.514 - Dec 14 16:31:09.527: INFO: Waiting up to 5m0s for pod "kube-proxy-mode-detector" in namespace "services-173" to be "running and ready" - Dec 14 16:31:09.532: INFO: Pod "kube-proxy-mode-detector": Phase="Pending", Reason="", readiness=false. Elapsed: 4.39727ms - Dec 14 16:31:09.532: INFO: The phase of Pod kube-proxy-mode-detector is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:31:11.541: INFO: Pod "kube-proxy-mode-detector": Phase="Running", Reason="", readiness=true. Elapsed: 2.013554727s - Dec 14 16:31:11.541: INFO: The phase of Pod kube-proxy-mode-detector is Running (Ready = true) - Dec 14 16:31:11.541: INFO: Pod "kube-proxy-mode-detector" satisfied condition "running and ready" - Dec 14 16:31:11.547: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-173 exec kube-proxy-mode-detector -- /bin/sh -x -c curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode' - Dec 14 16:31:11.851: INFO: stderr: "+ curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode\n" - Dec 14 16:31:11.851: INFO: stdout: "iptables" - Dec 14 16:31:11.851: INFO: proxyMode: iptables - Dec 14 16:31:11.872: INFO: Waiting for pod kube-proxy-mode-detector to disappear - Dec 14 16:31:11.876: INFO: Pod kube-proxy-mode-detector no longer exists - STEP: creating service affinity-clusterip-timeout in namespace services-173 12/14/22 16:31:11.876 - STEP: creating replication controller affinity-clusterip-timeout in namespace services-173 12/14/22 16:31:11.894 - I1214 16:31:11.906787 14 runners.go:193] Created replication controller with name: affinity-clusterip-timeout, namespace: services-173, replica count: 3 - I1214 16:31:14.957568 14 runners.go:193] affinity-clusterip-timeout Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - Dec 14 16:31:14.992: INFO: Creating new exec pod - Dec 14 16:31:15.004: INFO: Waiting up to 5m0s for pod "execpod-affinity7vgc4" in namespace "services-173" to be "running" - Dec 14 16:31:15.011: INFO: Pod "execpod-affinity7vgc4": Phase="Pending", Reason="", readiness=false. Elapsed: 7.027481ms - Dec 14 16:31:17.017: INFO: Pod "execpod-affinity7vgc4": Phase="Running", Reason="", readiness=true. Elapsed: 2.012640637s - Dec 14 16:31:17.017: INFO: Pod "execpod-affinity7vgc4" satisfied condition "running" - Dec 14 16:31:18.018: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-173 exec execpod-affinity7vgc4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-clusterip-timeout 80' - Dec 14 16:31:18.320: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-clusterip-timeout 80\nConnection to affinity-clusterip-timeout 80 port [tcp/http] succeeded!\n" - Dec 14 16:31:18.320: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:31:18.321: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-173 exec execpod-affinity7vgc4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.44.168 80' - Dec 14 16:31:18.590: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.44.168 80\nConnection to 10.233.44.168 80 port [tcp/http] succeeded!\n" - Dec 14 16:31:18.590: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:31:18.592: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-173 exec execpod-affinity7vgc4 -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.44.168:80/ ; done' - Dec 14 16:31:19.035: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n" - Dec 14 16:31:19.035: INFO: stdout: "\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4\naffinity-clusterip-timeout-qlpt4" - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Received response from host: affinity-clusterip-timeout-qlpt4 - Dec 14 16:31:19.035: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-173 exec execpod-affinity7vgc4 -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://10.233.44.168:80/' - Dec 14 16:31:19.243: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n" - Dec 14 16:31:19.243: INFO: stdout: "affinity-clusterip-timeout-qlpt4" - Dec 14 16:31:39.244: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-173 exec execpod-affinity7vgc4 -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://10.233.44.168:80/' - Dec 14 16:31:39.483: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://10.233.44.168:80/\n" - Dec 14 16:31:39.483: INFO: stdout: "affinity-clusterip-timeout-444jf" - Dec 14 16:31:39.483: INFO: Cleaning up the exec pod - STEP: deleting ReplicationController affinity-clusterip-timeout in namespace services-173, will wait for the garbage collector to delete the pods 12/14/22 16:31:39.508 - Dec 14 16:31:39.583: INFO: Deleting ReplicationController affinity-clusterip-timeout took: 13.564809ms - Dec 14 16:31:39.683: INFO: Terminating ReplicationController affinity-clusterip-timeout pods took: 100.496326ms + [It] should provide secure master service [Conformance] + test/e2e/network/service.go:781 [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 16:31:42.044: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-173" for this suite. 12/14/22 16:31:42.054 + Feb 25 10:34:32.653: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-3591" for this suite. 02/25/23 10:34:32.662 [AfterEach] [sig-network] Services test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -SS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-apps] Daemon set [Serial] - should run and stop simple daemon [Conformance] - test/e2e/apps/daemon_set.go:165 -[BeforeEach] [sig-apps] Daemon set [Serial] +[sig-api-machinery] ResourceQuota + should create a ResourceQuota and capture the life of a secret. [Conformance] + test/e2e/apimachinery/resource_quota.go:150 +[BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:31:42.064 -Dec 14 16:31:42.064: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename daemonsets 12/14/22 16:31:42.067 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:31:42.096 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:31:42.099 -[BeforeEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:145 -[It] should run and stop simple daemon [Conformance] - test/e2e/apps/daemon_set.go:165 -STEP: Creating simple DaemonSet "daemon-set" 12/14/22 16:31:42.14 -STEP: Check that daemon pods launch on every node of the cluster. 12/14/22 16:31:42.15 -Dec 14 16:31:42.161: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:31:42.162: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 16:31:43.184: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:31:43.184: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 16:31:44.186: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 -Dec 14 16:31:44.186: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set -STEP: Stop a daemon pod, check that the daemon pod is revived. 12/14/22 16:31:44.194 -Dec 14 16:31:44.227: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 -Dec 14 16:31:44.227: INFO: Node kiedai4iakae-2 is running 0 daemon pod, expected 1 -Dec 14 16:31:45.253: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 -Dec 14 16:31:45.253: INFO: Node kiedai4iakae-2 is running 0 daemon pod, expected 1 -Dec 14 16:31:46.245: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 -Dec 14 16:31:46.245: INFO: Node kiedai4iakae-2 is running 0 daemon pod, expected 1 -Dec 14 16:31:47.244: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 -Dec 14 16:31:47.244: INFO: Node kiedai4iakae-2 is running 0 daemon pod, expected 1 -Dec 14 16:31:48.241: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 -Dec 14 16:31:48.241: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set -[AfterEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:110 -STEP: Deleting DaemonSet "daemon-set" 12/14/22 16:31:48.246 -STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-83, will wait for the garbage collector to delete the pods 12/14/22 16:31:48.246 -Dec 14 16:31:48.310: INFO: Deleting DaemonSet.extensions daemon-set took: 9.396395ms -Dec 14 16:31:48.412: INFO: Terminating DaemonSet.extensions daemon-set pods took: 101.294895ms -Dec 14 16:31:50.417: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:31:50.418: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set -Dec 14 16:31:50.423: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"18647"},"items":null} - -Dec 14 16:31:50.427: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"18647"},"items":null} +STEP: Creating a kubernetes client 02/25/23 10:34:32.681 +Feb 25 10:34:32.682: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename resourcequota 02/25/23 10:34:32.685 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:34:32.711 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:34:32.721 +[It] should create a ResourceQuota and capture the life of a secret. [Conformance] + test/e2e/apimachinery/resource_quota.go:150 +STEP: Discovering how many secrets are in namespace by default 02/25/23 10:34:32.727 +STEP: Counting existing ResourceQuota 02/25/23 10:34:37.739 +STEP: Creating a ResourceQuota 02/25/23 10:34:42.747 +STEP: Ensuring resource quota status is calculated 02/25/23 10:34:42.769 +STEP: Creating a Secret 02/25/23 10:34:44.78 +STEP: Ensuring resource quota status captures secret creation 02/25/23 10:34:44.803 +STEP: Deleting a secret 02/25/23 10:34:46.813 +STEP: Ensuring resource quota status released usage 02/25/23 10:34:46.828 +[AfterEach] [sig-api-machinery] ResourceQuota + test/e2e/framework/framework.go:187 +Feb 25 10:34:48.842: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-9899" for this suite. 02/25/23 10:34:48.857 +{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a secret. [Conformance]","completed":154,"skipped":2807,"failed":0} +------------------------------ +• [SLOW TEST] [16.191 seconds] +[sig-api-machinery] ResourceQuota +test/e2e/apimachinery/framework.go:23 + should create a ResourceQuota and capture the life of a secret. [Conformance] + test/e2e/apimachinery/resource_quota.go:150 -[AfterEach] [sig-apps] Daemon set [Serial] + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-api-machinery] ResourceQuota + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:34:32.681 + Feb 25 10:34:32.682: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename resourcequota 02/25/23 10:34:32.685 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:34:32.711 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:34:32.721 + [It] should create a ResourceQuota and capture the life of a secret. [Conformance] + test/e2e/apimachinery/resource_quota.go:150 + STEP: Discovering how many secrets are in namespace by default 02/25/23 10:34:32.727 + STEP: Counting existing ResourceQuota 02/25/23 10:34:37.739 + STEP: Creating a ResourceQuota 02/25/23 10:34:42.747 + STEP: Ensuring resource quota status is calculated 02/25/23 10:34:42.769 + STEP: Creating a Secret 02/25/23 10:34:44.78 + STEP: Ensuring resource quota status captures secret creation 02/25/23 10:34:44.803 + STEP: Deleting a secret 02/25/23 10:34:46.813 + STEP: Ensuring resource quota status released usage 02/25/23 10:34:46.828 + [AfterEach] [sig-api-machinery] ResourceQuota + test/e2e/framework/framework.go:187 + Feb 25 10:34:48.842: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "resourcequota-9899" for this suite. 02/25/23 10:34:48.857 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-apps] ReplicationController + should surface a failure condition on a common issue like exceeded quota [Conformance] + test/e2e/apps/rc.go:82 +[BeforeEach] [sig-apps] ReplicationController + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:34:48.877 +Feb 25 10:34:48.878: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename replication-controller 02/25/23 10:34:48.883 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:34:48.993 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:34:49 +[BeforeEach] [sig-apps] ReplicationController + test/e2e/apps/rc.go:56 +[It] should surface a failure condition on a common issue like exceeded quota [Conformance] + test/e2e/apps/rc.go:82 +Feb 25 10:34:49.017: INFO: Creating quota "condition-test" that allows only two pods to run in the current namespace +STEP: Creating rc "condition-test" that asks for more than the allowed pod quota 02/25/23 10:34:50.082 +STEP: Checking rc "condition-test" has the desired failure condition set 02/25/23 10:34:50.1 +STEP: Scaling down rc "condition-test" to satisfy pod quota 02/25/23 10:34:51.123 +Feb 25 10:34:51.151: INFO: Updating replication controller "condition-test" +STEP: Checking rc "condition-test" has no failure condition set 02/25/23 10:34:51.151 +[AfterEach] [sig-apps] ReplicationController test/e2e/framework/framework.go:187 -Dec 14 16:31:50.453: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "daemonsets-83" for this suite. 12/14/22 16:31:50.46 -{"msg":"PASSED [sig-apps] Daemon set [Serial] should run and stop simple daemon [Conformance]","completed":160,"skipped":2876,"failed":0} +Feb 25 10:34:52.173: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "replication-controller-4268" for this suite. 02/25/23 10:34:52.187 +{"msg":"PASSED [sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance]","completed":155,"skipped":2830,"failed":0} ------------------------------ -• [SLOW TEST] [8.415 seconds] -[sig-apps] Daemon set [Serial] +• [3.328 seconds] +[sig-apps] ReplicationController test/e2e/apps/framework.go:23 - should run and stop simple daemon [Conformance] - test/e2e/apps/daemon_set.go:165 + should surface a failure condition on a common issue like exceeded quota [Conformance] + test/e2e/apps/rc.go:82 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Daemon set [Serial] + [BeforeEach] [sig-apps] ReplicationController test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:31:42.064 - Dec 14 16:31:42.064: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename daemonsets 12/14/22 16:31:42.067 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:31:42.096 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:31:42.099 - [BeforeEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:145 - [It] should run and stop simple daemon [Conformance] - test/e2e/apps/daemon_set.go:165 - STEP: Creating simple DaemonSet "daemon-set" 12/14/22 16:31:42.14 - STEP: Check that daemon pods launch on every node of the cluster. 12/14/22 16:31:42.15 - Dec 14 16:31:42.161: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:31:42.162: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 16:31:43.184: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:31:43.184: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 16:31:44.186: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 - Dec 14 16:31:44.186: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set - STEP: Stop a daemon pod, check that the daemon pod is revived. 12/14/22 16:31:44.194 - Dec 14 16:31:44.227: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 - Dec 14 16:31:44.227: INFO: Node kiedai4iakae-2 is running 0 daemon pod, expected 1 - Dec 14 16:31:45.253: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 - Dec 14 16:31:45.253: INFO: Node kiedai4iakae-2 is running 0 daemon pod, expected 1 - Dec 14 16:31:46.245: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 - Dec 14 16:31:46.245: INFO: Node kiedai4iakae-2 is running 0 daemon pod, expected 1 - Dec 14 16:31:47.244: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 - Dec 14 16:31:47.244: INFO: Node kiedai4iakae-2 is running 0 daemon pod, expected 1 - Dec 14 16:31:48.241: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 - Dec 14 16:31:48.241: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set - [AfterEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:110 - STEP: Deleting DaemonSet "daemon-set" 12/14/22 16:31:48.246 - STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-83, will wait for the garbage collector to delete the pods 12/14/22 16:31:48.246 - Dec 14 16:31:48.310: INFO: Deleting DaemonSet.extensions daemon-set took: 9.396395ms - Dec 14 16:31:48.412: INFO: Terminating DaemonSet.extensions daemon-set pods took: 101.294895ms - Dec 14 16:31:50.417: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:31:50.418: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set - Dec 14 16:31:50.423: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"18647"},"items":null} - - Dec 14 16:31:50.427: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"18647"},"items":null} + STEP: Creating a kubernetes client 02/25/23 10:34:48.877 + Feb 25 10:34:48.878: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename replication-controller 02/25/23 10:34:48.883 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:34:48.993 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:34:49 + [BeforeEach] [sig-apps] ReplicationController + test/e2e/apps/rc.go:56 + [It] should surface a failure condition on a common issue like exceeded quota [Conformance] + test/e2e/apps/rc.go:82 + Feb 25 10:34:49.017: INFO: Creating quota "condition-test" that allows only two pods to run in the current namespace + STEP: Creating rc "condition-test" that asks for more than the allowed pod quota 02/25/23 10:34:50.082 + STEP: Checking rc "condition-test" has the desired failure condition set 02/25/23 10:34:50.1 + STEP: Scaling down rc "condition-test" to satisfy pod quota 02/25/23 10:34:51.123 + Feb 25 10:34:51.151: INFO: Updating replication controller "condition-test" + STEP: Checking rc "condition-test" has no failure condition set 02/25/23 10:34:51.151 + [AfterEach] [sig-apps] ReplicationController + test/e2e/framework/framework.go:187 + Feb 25 10:34:52.173: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "replication-controller-4268" for this suite. 02/25/23 10:34:52.187 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook + should execute prestop exec hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:114 +[BeforeEach] [sig-node] Container Lifecycle Hook + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:34:52.212 +Feb 25 10:34:52.212: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-lifecycle-hook 02/25/23 10:34:52.217 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:34:52.273 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:34:52.278 +[BeforeEach] when create a pod with lifecycle hook + test/e2e/common/node/lifecycle_hook.go:55 +STEP: create the container to handle the HTTPGet hook request. 02/25/23 10:34:52.295 +Feb 25 10:34:52.314: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-3047" to be "running and ready" +Feb 25 10:34:52.321: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 6.787832ms +Feb 25 10:34:52.321: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:34:54.330: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016081535s +Feb 25 10:34:54.330: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:34:56.328: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 4.014397908s +Feb 25 10:34:56.328: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) +Feb 25 10:34:56.328: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" +[It] should execute prestop exec hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:114 +STEP: create the pod with lifecycle hook 02/25/23 10:34:56.334 +Feb 25 10:34:56.345: INFO: Waiting up to 5m0s for pod "pod-with-prestop-exec-hook" in namespace "container-lifecycle-hook-3047" to be "running and ready" +Feb 25 10:34:56.362: INFO: Pod "pod-with-prestop-exec-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 16.381855ms +Feb 25 10:34:56.362: INFO: The phase of Pod pod-with-prestop-exec-hook is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:34:58.376: INFO: Pod "pod-with-prestop-exec-hook": Phase="Running", Reason="", readiness=true. Elapsed: 2.030566846s +Feb 25 10:34:58.376: INFO: The phase of Pod pod-with-prestop-exec-hook is Running (Ready = true) +Feb 25 10:34:58.376: INFO: Pod "pod-with-prestop-exec-hook" satisfied condition "running and ready" +STEP: delete the pod with lifecycle hook 02/25/23 10:34:58.39 +Feb 25 10:34:58.422: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Feb 25 10:34:58.433: INFO: Pod pod-with-prestop-exec-hook still exists +Feb 25 10:35:00.433: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Feb 25 10:35:00.442: INFO: Pod pod-with-prestop-exec-hook still exists +Feb 25 10:35:02.433: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear +Feb 25 10:35:02.442: INFO: Pod pod-with-prestop-exec-hook no longer exists +STEP: check prestop hook 02/25/23 10:35:02.442 +[AfterEach] [sig-node] Container Lifecycle Hook + test/e2e/framework/framework.go:187 +Feb 25 10:35:02.496: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-lifecycle-hook-3047" for this suite. 02/25/23 10:35:02.508 +{"msg":"PASSED [sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance]","completed":156,"skipped":2864,"failed":0} +------------------------------ +• [SLOW TEST] [10.319 seconds] +[sig-node] Container Lifecycle Hook +test/e2e/common/node/framework.go:23 + when create a pod with lifecycle hook + test/e2e/common/node/lifecycle_hook.go:46 + should execute prestop exec hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:114 - [AfterEach] [sig-apps] Daemon set [Serial] + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-node] Container Lifecycle Hook + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:34:52.212 + Feb 25 10:34:52.212: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-lifecycle-hook 02/25/23 10:34:52.217 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:34:52.273 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:34:52.278 + [BeforeEach] when create a pod with lifecycle hook + test/e2e/common/node/lifecycle_hook.go:55 + STEP: create the container to handle the HTTPGet hook request. 02/25/23 10:34:52.295 + Feb 25 10:34:52.314: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-3047" to be "running and ready" + Feb 25 10:34:52.321: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 6.787832ms + Feb 25 10:34:52.321: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:34:54.330: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016081535s + Feb 25 10:34:54.330: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:34:56.328: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 4.014397908s + Feb 25 10:34:56.328: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) + Feb 25 10:34:56.328: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" + [It] should execute prestop exec hook properly [NodeConformance] [Conformance] + test/e2e/common/node/lifecycle_hook.go:114 + STEP: create the pod with lifecycle hook 02/25/23 10:34:56.334 + Feb 25 10:34:56.345: INFO: Waiting up to 5m0s for pod "pod-with-prestop-exec-hook" in namespace "container-lifecycle-hook-3047" to be "running and ready" + Feb 25 10:34:56.362: INFO: Pod "pod-with-prestop-exec-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 16.381855ms + Feb 25 10:34:56.362: INFO: The phase of Pod pod-with-prestop-exec-hook is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:34:58.376: INFO: Pod "pod-with-prestop-exec-hook": Phase="Running", Reason="", readiness=true. Elapsed: 2.030566846s + Feb 25 10:34:58.376: INFO: The phase of Pod pod-with-prestop-exec-hook is Running (Ready = true) + Feb 25 10:34:58.376: INFO: Pod "pod-with-prestop-exec-hook" satisfied condition "running and ready" + STEP: delete the pod with lifecycle hook 02/25/23 10:34:58.39 + Feb 25 10:34:58.422: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear + Feb 25 10:34:58.433: INFO: Pod pod-with-prestop-exec-hook still exists + Feb 25 10:35:00.433: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear + Feb 25 10:35:00.442: INFO: Pod pod-with-prestop-exec-hook still exists + Feb 25 10:35:02.433: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear + Feb 25 10:35:02.442: INFO: Pod pod-with-prestop-exec-hook no longer exists + STEP: check prestop hook 02/25/23 10:35:02.442 + [AfterEach] [sig-node] Container Lifecycle Hook test/e2e/framework/framework.go:187 - Dec 14 16:31:50.453: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "daemonsets-83" for this suite. 12/14/22 16:31:50.46 + Feb 25 10:35:02.496: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-lifecycle-hook-3047" for this suite. 02/25/23 10:35:02.508 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-apps] DisruptionController - should observe PodDisruptionBudget status updated [Conformance] - test/e2e/apps/disruption.go:140 -[BeforeEach] [sig-apps] DisruptionController +[sig-node] Kubelet when scheduling a busybox command that always fails in a pod + should be possible to delete [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:135 +[BeforeEach] [sig-node] Kubelet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:31:50.484 -Dec 14 16:31:50.484: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename disruption 12/14/22 16:31:50.491 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:31:50.522 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:31:50.527 -[BeforeEach] [sig-apps] DisruptionController - test/e2e/apps/disruption.go:71 -[It] should observe PodDisruptionBudget status updated [Conformance] - test/e2e/apps/disruption.go:140 -STEP: Waiting for the pdb to be processed 12/14/22 16:31:50.538 -STEP: Waiting for all pods to be running 12/14/22 16:31:52.593 -Dec 14 16:31:52.610: INFO: running pods: 0 < 3 -[AfterEach] [sig-apps] DisruptionController +STEP: Creating a kubernetes client 02/25/23 10:35:02.549 +Feb 25 10:35:02.549: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubelet-test 02/25/23 10:35:02.551 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:35:02.586 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:35:02.597 +[BeforeEach] [sig-node] Kubelet + test/e2e/common/node/kubelet.go:41 +[BeforeEach] when scheduling a busybox command that always fails in a pod + test/e2e/common/node/kubelet.go:85 +[It] should be possible to delete [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:135 +[AfterEach] [sig-node] Kubelet test/e2e/framework/framework.go:187 -Dec 14 16:31:54.633: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "disruption-3399" for this suite. 12/14/22 16:31:54.64 -{"msg":"PASSED [sig-apps] DisruptionController should observe PodDisruptionBudget status updated [Conformance]","completed":161,"skipped":2879,"failed":0} +Feb 25 10:35:02.650: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubelet-test-6337" for this suite. 02/25/23 10:35:02.668 +{"msg":"PASSED [sig-node] Kubelet when scheduling a busybox command that always fails in a pod should be possible to delete [NodeConformance] [Conformance]","completed":157,"skipped":2914,"failed":0} ------------------------------ -• [4.166 seconds] -[sig-apps] DisruptionController -test/e2e/apps/framework.go:23 - should observe PodDisruptionBudget status updated [Conformance] - test/e2e/apps/disruption.go:140 +• [0.148 seconds] +[sig-node] Kubelet +test/e2e/common/node/framework.go:23 + when scheduling a busybox command that always fails in a pod + test/e2e/common/node/kubelet.go:82 + should be possible to delete [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:135 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] DisruptionController + [BeforeEach] [sig-node] Kubelet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:31:50.484 - Dec 14 16:31:50.484: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename disruption 12/14/22 16:31:50.491 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:31:50.522 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:31:50.527 - [BeforeEach] [sig-apps] DisruptionController - test/e2e/apps/disruption.go:71 - [It] should observe PodDisruptionBudget status updated [Conformance] - test/e2e/apps/disruption.go:140 - STEP: Waiting for the pdb to be processed 12/14/22 16:31:50.538 - STEP: Waiting for all pods to be running 12/14/22 16:31:52.593 - Dec 14 16:31:52.610: INFO: running pods: 0 < 3 - [AfterEach] [sig-apps] DisruptionController + STEP: Creating a kubernetes client 02/25/23 10:35:02.549 + Feb 25 10:35:02.549: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubelet-test 02/25/23 10:35:02.551 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:35:02.586 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:35:02.597 + [BeforeEach] [sig-node] Kubelet + test/e2e/common/node/kubelet.go:41 + [BeforeEach] when scheduling a busybox command that always fails in a pod + test/e2e/common/node/kubelet.go:85 + [It] should be possible to delete [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:135 + [AfterEach] [sig-node] Kubelet test/e2e/framework/framework.go:187 - Dec 14 16:31:54.633: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "disruption-3399" for this suite. 12/14/22 16:31:54.64 + Feb 25 10:35:02.650: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubelet-test-6337" for this suite. 02/25/23 10:35:02.668 << End Captured GinkgoWriter Output ------------------------------ S ------------------------------ -[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] - should list, patch and delete a collection of StatefulSets [Conformance] - test/e2e/apps/statefulset.go:906 -[BeforeEach] [sig-apps] StatefulSet +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + should include webhook resources in discovery documents [Conformance] + test/e2e/apimachinery/webhook.go:116 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:31:54.651 -Dec 14 16:31:54.652: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename statefulset 12/14/22 16:31:54.654 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:31:54.679 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:31:54.682 -[BeforeEach] [sig-apps] StatefulSet - test/e2e/apps/statefulset.go:96 -[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:111 -STEP: Creating service test in namespace statefulset-2566 12/14/22 16:31:54.688 -[It] should list, patch and delete a collection of StatefulSets [Conformance] - test/e2e/apps/statefulset.go:906 -Dec 14 16:31:54.708: INFO: Found 0 stateful pods, waiting for 1 -Dec 14 16:32:04.723: INFO: Waiting for pod test-ss-0 to enter Running - Ready=true, currently Running - Ready=true -STEP: patching the StatefulSet 12/14/22 16:32:04.738 -W1214 16:32:04.766245 14 warnings.go:70] unknown field "spec.template.spec.TerminationGracePeriodSeconds" -Dec 14 16:32:04.780: INFO: Found 1 stateful pods, waiting for 2 -Dec 14 16:32:14.792: INFO: Waiting for pod test-ss-0 to enter Running - Ready=true, currently Running - Ready=true -Dec 14 16:32:14.792: INFO: Waiting for pod test-ss-1 to enter Running - Ready=true, currently Running - Ready=true -STEP: Listing all StatefulSets 12/14/22 16:32:14.806 -STEP: Delete all of the StatefulSets 12/14/22 16:32:14.813 -STEP: Verify that StatefulSets have been deleted 12/14/22 16:32:14.829 -[AfterEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:122 -Dec 14 16:32:14.834: INFO: Deleting all statefulset in ns statefulset-2566 -[AfterEach] [sig-apps] StatefulSet +STEP: Creating a kubernetes client 02/25/23 10:35:02.698 +Feb 25 10:35:02.698: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 10:35:02.705 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:35:02.742 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:35:02.751 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 +STEP: Setting up server cert 02/25/23 10:35:02.803 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:35:04.42 +STEP: Deploying the webhook pod 02/25/23 10:35:04.434 +STEP: Wait for the deployment to be ready 02/25/23 10:35:04.457 +Feb 25 10:35:04.475: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created +STEP: Deploying the webhook service 02/25/23 10:35:06.496 +STEP: Verifying the service has paired with the endpoint 02/25/23 10:35:06.529 +Feb 25 10:35:07.530: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should include webhook resources in discovery documents [Conformance] + test/e2e/apimachinery/webhook.go:116 +STEP: fetching the /apis discovery document 02/25/23 10:35:07.537 +STEP: finding the admissionregistration.k8s.io API group in the /apis discovery document 02/25/23 10:35:07.544 +STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis discovery document 02/25/23 10:35:07.544 +STEP: fetching the /apis/admissionregistration.k8s.io discovery document 02/25/23 10:35:07.544 +STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis/admissionregistration.k8s.io discovery document 02/25/23 10:35:07.552 +STEP: fetching the /apis/admissionregistration.k8s.io/v1 discovery document 02/25/23 10:35:07.552 +STEP: finding mutatingwebhookconfigurations and validatingwebhookconfigurations resources in the /apis/admissionregistration.k8s.io/v1 discovery document 02/25/23 10:35:07.555 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:32:14.849: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "statefulset-2566" for this suite. 12/14/22 16:32:14.857 -{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should list, patch and delete a collection of StatefulSets [Conformance]","completed":162,"skipped":2880,"failed":0} +Feb 25 10:35:07.555: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-6552" for this suite. 02/25/23 10:35:07.564 +STEP: Destroying namespace "webhook-6552-markers" for this suite. 02/25/23 10:35:07.59 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should include webhook resources in discovery documents [Conformance]","completed":158,"skipped":2915,"failed":0} ------------------------------ -• [SLOW TEST] [20.246 seconds] -[sig-apps] StatefulSet -test/e2e/apps/framework.go:23 - Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:101 - should list, patch and delete a collection of StatefulSets [Conformance] - test/e2e/apps/statefulset.go:906 +• [SLOW TEST] [5.086 seconds] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + should include webhook resources in discovery documents [Conformance] + test/e2e/apimachinery/webhook.go:116 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:35:02.698 + Feb 25 10:35:02.698: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 10:35:02.705 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:35:02.742 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:35:02.751 + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 + STEP: Setting up server cert 02/25/23 10:35:02.803 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:35:04.42 + STEP: Deploying the webhook pod 02/25/23 10:35:04.434 + STEP: Wait for the deployment to be ready 02/25/23 10:35:04.457 + Feb 25 10:35:04.475: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created + STEP: Deploying the webhook service 02/25/23 10:35:06.496 + STEP: Verifying the service has paired with the endpoint 02/25/23 10:35:06.529 + Feb 25 10:35:07.530: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] should include webhook resources in discovery documents [Conformance] + test/e2e/apimachinery/webhook.go:116 + STEP: fetching the /apis discovery document 02/25/23 10:35:07.537 + STEP: finding the admissionregistration.k8s.io API group in the /apis discovery document 02/25/23 10:35:07.544 + STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis discovery document 02/25/23 10:35:07.544 + STEP: fetching the /apis/admissionregistration.k8s.io discovery document 02/25/23 10:35:07.544 + STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis/admissionregistration.k8s.io discovery document 02/25/23 10:35:07.552 + STEP: fetching the /apis/admissionregistration.k8s.io/v1 discovery document 02/25/23 10:35:07.552 + STEP: finding mutatingwebhookconfigurations and validatingwebhookconfigurations resources in the /apis/admissionregistration.k8s.io/v1 discovery document 02/25/23 10:35:07.555 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/framework/framework.go:187 + Feb 25 10:35:07.555: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-6552" for this suite. 02/25/23 10:35:07.564 + STEP: Destroying namespace "webhook-6552-markers" for this suite. 02/25/23 10:35:07.59 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-node] ConfigMap + should be consumable via environment variable [NodeConformance] [Conformance] + test/e2e/common/node/configmap.go:44 +[BeforeEach] [sig-node] ConfigMap + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:35:07.838 +Feb 25 10:35:07.841: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename configmap 02/25/23 10:35:07.844 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:35:07.905 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:35:07.919 +[It] should be consumable via environment variable [NodeConformance] [Conformance] + test/e2e/common/node/configmap.go:44 +STEP: Creating configMap configmap-347/configmap-test-9c2ce510-9ba1-4dd7-87f8-6fe6cf5692e3 02/25/23 10:35:07.923 +STEP: Creating a pod to test consume configMaps 02/25/23 10:35:07.937 +Feb 25 10:35:07.971: INFO: Waiting up to 5m0s for pod "pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7" in namespace "configmap-347" to be "Succeeded or Failed" +Feb 25 10:35:07.983: INFO: Pod "pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7": Phase="Pending", Reason="", readiness=false. Elapsed: 11.900307ms +Feb 25 10:35:09.992: INFO: Pod "pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020969998s +Feb 25 10:35:11.994: INFO: Pod "pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022028312s +STEP: Saw pod success 02/25/23 10:35:11.994 +Feb 25 10:35:11.994: INFO: Pod "pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7" satisfied condition "Succeeded or Failed" +Feb 25 10:35:12.000: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7 container env-test: +STEP: delete the pod 02/25/23 10:35:12.042 +Feb 25 10:35:12.077: INFO: Waiting for pod pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7 to disappear +Feb 25 10:35:12.088: INFO: Pod pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7 no longer exists +[AfterEach] [sig-node] ConfigMap + test/e2e/framework/framework.go:187 +Feb 25 10:35:12.088: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-347" for this suite. 02/25/23 10:35:12.096 +{"msg":"PASSED [sig-node] ConfigMap should be consumable via environment variable [NodeConformance] [Conformance]","completed":159,"skipped":2946,"failed":0} +------------------------------ +• [4.281 seconds] +[sig-node] ConfigMap +test/e2e/common/node/framework.go:23 + should be consumable via environment variable [NodeConformance] [Conformance] + test/e2e/common/node/configmap.go:44 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-node] ConfigMap + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:35:07.838 + Feb 25 10:35:07.841: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename configmap 02/25/23 10:35:07.844 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:35:07.905 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:35:07.919 + [It] should be consumable via environment variable [NodeConformance] [Conformance] + test/e2e/common/node/configmap.go:44 + STEP: Creating configMap configmap-347/configmap-test-9c2ce510-9ba1-4dd7-87f8-6fe6cf5692e3 02/25/23 10:35:07.923 + STEP: Creating a pod to test consume configMaps 02/25/23 10:35:07.937 + Feb 25 10:35:07.971: INFO: Waiting up to 5m0s for pod "pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7" in namespace "configmap-347" to be "Succeeded or Failed" + Feb 25 10:35:07.983: INFO: Pod "pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7": Phase="Pending", Reason="", readiness=false. Elapsed: 11.900307ms + Feb 25 10:35:09.992: INFO: Pod "pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020969998s + Feb 25 10:35:11.994: INFO: Pod "pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022028312s + STEP: Saw pod success 02/25/23 10:35:11.994 + Feb 25 10:35:11.994: INFO: Pod "pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7" satisfied condition "Succeeded or Failed" + Feb 25 10:35:12.000: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7 container env-test: + STEP: delete the pod 02/25/23 10:35:12.042 + Feb 25 10:35:12.077: INFO: Waiting for pod pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7 to disappear + Feb 25 10:35:12.088: INFO: Pod pod-configmaps-7a19e525-1cf2-4bfa-b69d-101632b324f7 no longer exists + [AfterEach] [sig-node] ConfigMap + test/e2e/framework/framework.go:187 + Feb 25 10:35:12.088: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "configmap-347" for this suite. 02/25/23 10:35:12.096 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSS +------------------------------ +[sig-storage] Projected downwardAPI + should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:83 +[BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:35:12.124 +Feb 25 10:35:12.125: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 10:35:12.129 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:35:12.167 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:35:12.172 +[BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 +[It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:83 +STEP: Creating a pod to test downward API volume plugin 02/25/23 10:35:12.181 +Feb 25 10:35:12.206: INFO: Waiting up to 5m0s for pod "downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd" in namespace "projected-3506" to be "Succeeded or Failed" +Feb 25 10:35:12.216: INFO: Pod "downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd": Phase="Pending", Reason="", readiness=false. Elapsed: 10.166207ms +Feb 25 10:35:14.224: INFO: Pod "downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018119281s +Feb 25 10:35:16.232: INFO: Pod "downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd": Phase="Pending", Reason="", readiness=false. Elapsed: 4.025591288s +Feb 25 10:35:18.225: INFO: Pod "downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.01909919s +STEP: Saw pod success 02/25/23 10:35:18.229 +Feb 25 10:35:18.230: INFO: Pod "downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd" satisfied condition "Succeeded or Failed" +Feb 25 10:35:18.240: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd container client-container: +STEP: delete the pod 02/25/23 10:35:18.264 +Feb 25 10:35:18.292: INFO: Waiting for pod downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd to disappear +Feb 25 10:35:18.299: INFO: Pod downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd no longer exists +[AfterEach] [sig-storage] Projected downwardAPI + test/e2e/framework/framework.go:187 +Feb 25 10:35:18.300: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-3506" for this suite. 02/25/23 10:35:18.31 +{"msg":"PASSED [sig-storage] Projected downwardAPI should set mode on item file [LinuxOnly] [NodeConformance] [Conformance]","completed":160,"skipped":2957,"failed":0} +------------------------------ +• [SLOW TEST] [6.205 seconds] +[sig-storage] Projected downwardAPI +test/e2e/common/storage/framework.go:23 + should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:83 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] StatefulSet + [BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:31:54.651 - Dec 14 16:31:54.652: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename statefulset 12/14/22 16:31:54.654 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:31:54.679 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:31:54.682 - [BeforeEach] [sig-apps] StatefulSet - test/e2e/apps/statefulset.go:96 - [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:111 - STEP: Creating service test in namespace statefulset-2566 12/14/22 16:31:54.688 - [It] should list, patch and delete a collection of StatefulSets [Conformance] - test/e2e/apps/statefulset.go:906 - Dec 14 16:31:54.708: INFO: Found 0 stateful pods, waiting for 1 - Dec 14 16:32:04.723: INFO: Waiting for pod test-ss-0 to enter Running - Ready=true, currently Running - Ready=true - STEP: patching the StatefulSet 12/14/22 16:32:04.738 - W1214 16:32:04.766245 14 warnings.go:70] unknown field "spec.template.spec.TerminationGracePeriodSeconds" - Dec 14 16:32:04.780: INFO: Found 1 stateful pods, waiting for 2 - Dec 14 16:32:14.792: INFO: Waiting for pod test-ss-0 to enter Running - Ready=true, currently Running - Ready=true - Dec 14 16:32:14.792: INFO: Waiting for pod test-ss-1 to enter Running - Ready=true, currently Running - Ready=true - STEP: Listing all StatefulSets 12/14/22 16:32:14.806 - STEP: Delete all of the StatefulSets 12/14/22 16:32:14.813 - STEP: Verify that StatefulSets have been deleted 12/14/22 16:32:14.829 - [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:122 - Dec 14 16:32:14.834: INFO: Deleting all statefulset in ns statefulset-2566 - [AfterEach] [sig-apps] StatefulSet + STEP: Creating a kubernetes client 02/25/23 10:35:12.124 + Feb 25 10:35:12.125: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 10:35:12.129 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:35:12.167 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:35:12.172 + [BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 + [It] should set mode on item file [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:83 + STEP: Creating a pod to test downward API volume plugin 02/25/23 10:35:12.181 + Feb 25 10:35:12.206: INFO: Waiting up to 5m0s for pod "downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd" in namespace "projected-3506" to be "Succeeded or Failed" + Feb 25 10:35:12.216: INFO: Pod "downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd": Phase="Pending", Reason="", readiness=false. Elapsed: 10.166207ms + Feb 25 10:35:14.224: INFO: Pod "downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018119281s + Feb 25 10:35:16.232: INFO: Pod "downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd": Phase="Pending", Reason="", readiness=false. Elapsed: 4.025591288s + Feb 25 10:35:18.225: INFO: Pod "downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.01909919s + STEP: Saw pod success 02/25/23 10:35:18.229 + Feb 25 10:35:18.230: INFO: Pod "downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd" satisfied condition "Succeeded or Failed" + Feb 25 10:35:18.240: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd container client-container: + STEP: delete the pod 02/25/23 10:35:18.264 + Feb 25 10:35:18.292: INFO: Waiting for pod downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd to disappear + Feb 25 10:35:18.299: INFO: Pod downwardapi-volume-c8a2f1cd-3dc9-4f87-a9f0-f68aaba56fdd no longer exists + [AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 - Dec 14 16:32:14.849: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "statefulset-2566" for this suite. 12/14/22 16:32:14.857 + Feb 25 10:35:18.300: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-3506" for this suite. 02/25/23 10:35:18.31 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Secrets - should fail to create secret due to empty secret key [Conformance] - test/e2e/common/node/secrets.go:139 -[BeforeEach] [sig-node] Secrets +[sig-storage] Secrets + optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:204 +[BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:32:14.907 -Dec 14 16:32:14.907: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename secrets 12/14/22 16:32:14.91 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:32:14.938 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:32:14.941 -[It] should fail to create secret due to empty secret key [Conformance] - test/e2e/common/node/secrets.go:139 -STEP: Creating projection with secret that has name secret-emptykey-test-86b9f76c-87e0-4ea2-8c79-35390b142fe0 12/14/22 16:32:14.944 -[AfterEach] [sig-node] Secrets +STEP: Creating a kubernetes client 02/25/23 10:35:18.342 +Feb 25 10:35:18.343: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename secrets 02/25/23 10:35:18.347 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:35:18.389 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:35:18.397 +[It] optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:204 +STEP: Creating secret with name s-test-opt-del-8eaa75ae-6026-4c96-9160-1c250fab32f4 02/25/23 10:35:18.425 +STEP: Creating secret with name s-test-opt-upd-9f80df9c-0605-4ddb-bda5-c9f9f8ae016e 02/25/23 10:35:18.44 +STEP: Creating the pod 02/25/23 10:35:18.45 +Feb 25 10:35:18.477: INFO: Waiting up to 5m0s for pod "pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466" in namespace "secrets-7898" to be "running and ready" +Feb 25 10:35:18.488: INFO: Pod "pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466": Phase="Pending", Reason="", readiness=false. Elapsed: 10.165426ms +Feb 25 10:35:18.488: INFO: The phase of Pod pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:35:20.496: INFO: Pod "pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01871975s +Feb 25 10:35:20.496: INFO: The phase of Pod pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:35:22.512: INFO: Pod "pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466": Phase="Running", Reason="", readiness=true. Elapsed: 4.034205961s +Feb 25 10:35:22.512: INFO: The phase of Pod pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466 is Running (Ready = true) +Feb 25 10:35:22.512: INFO: Pod "pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466" satisfied condition "running and ready" +STEP: Deleting secret s-test-opt-del-8eaa75ae-6026-4c96-9160-1c250fab32f4 02/25/23 10:35:22.577 +STEP: Updating secret s-test-opt-upd-9f80df9c-0605-4ddb-bda5-c9f9f8ae016e 02/25/23 10:35:22.595 +STEP: Creating secret with name s-test-opt-create-a099c2e5-2441-4539-ac3d-fab9d82a69ca 02/25/23 10:35:22.61 +STEP: waiting to observe update in volume 02/25/23 10:35:22.625 +[AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 -Dec 14 16:32:14.947: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-4462" for this suite. 12/14/22 16:32:14.956 -{"msg":"PASSED [sig-node] Secrets should fail to create secret due to empty secret key [Conformance]","completed":163,"skipped":2901,"failed":0} +Feb 25 10:36:42.086: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-7898" for this suite. 02/25/23 10:36:42.107 +{"msg":"PASSED [sig-storage] Secrets optional updates should be reflected in volume [NodeConformance] [Conformance]","completed":161,"skipped":2987,"failed":0} ------------------------------ -• [0.060 seconds] -[sig-node] Secrets -test/e2e/common/node/framework.go:23 - should fail to create secret due to empty secret key [Conformance] - test/e2e/common/node/secrets.go:139 +• [SLOW TEST] [83.789 seconds] +[sig-storage] Secrets +test/e2e/common/storage/framework.go:23 + optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:204 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Secrets + [BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:32:14.907 - Dec 14 16:32:14.907: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename secrets 12/14/22 16:32:14.91 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:32:14.938 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:32:14.941 - [It] should fail to create secret due to empty secret key [Conformance] - test/e2e/common/node/secrets.go:139 - STEP: Creating projection with secret that has name secret-emptykey-test-86b9f76c-87e0-4ea2-8c79-35390b142fe0 12/14/22 16:32:14.944 - [AfterEach] [sig-node] Secrets + STEP: Creating a kubernetes client 02/25/23 10:35:18.342 + Feb 25 10:35:18.343: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename secrets 02/25/23 10:35:18.347 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:35:18.389 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:35:18.397 + [It] optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:204 + STEP: Creating secret with name s-test-opt-del-8eaa75ae-6026-4c96-9160-1c250fab32f4 02/25/23 10:35:18.425 + STEP: Creating secret with name s-test-opt-upd-9f80df9c-0605-4ddb-bda5-c9f9f8ae016e 02/25/23 10:35:18.44 + STEP: Creating the pod 02/25/23 10:35:18.45 + Feb 25 10:35:18.477: INFO: Waiting up to 5m0s for pod "pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466" in namespace "secrets-7898" to be "running and ready" + Feb 25 10:35:18.488: INFO: Pod "pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466": Phase="Pending", Reason="", readiness=false. Elapsed: 10.165426ms + Feb 25 10:35:18.488: INFO: The phase of Pod pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:35:20.496: INFO: Pod "pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01871975s + Feb 25 10:35:20.496: INFO: The phase of Pod pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:35:22.512: INFO: Pod "pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466": Phase="Running", Reason="", readiness=true. Elapsed: 4.034205961s + Feb 25 10:35:22.512: INFO: The phase of Pod pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466 is Running (Ready = true) + Feb 25 10:35:22.512: INFO: Pod "pod-secrets-fc8b47b3-0f1c-4f7f-8884-ab95cec01466" satisfied condition "running and ready" + STEP: Deleting secret s-test-opt-del-8eaa75ae-6026-4c96-9160-1c250fab32f4 02/25/23 10:35:22.577 + STEP: Updating secret s-test-opt-upd-9f80df9c-0605-4ddb-bda5-c9f9f8ae016e 02/25/23 10:35:22.595 + STEP: Creating secret with name s-test-opt-create-a099c2e5-2441-4539-ac3d-fab9d82a69ca 02/25/23 10:35:22.61 + STEP: waiting to observe update in volume 02/25/23 10:35:22.625 + [AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 - Dec 14 16:32:14.947: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "secrets-4462" for this suite. 12/14/22 16:32:14.956 + Feb 25 10:36:42.086: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "secrets-7898" for this suite. 02/25/23 10:36:42.107 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Projected configMap - should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:108 -[BeforeEach] [sig-storage] Projected configMap +[sig-cli] Kubectl client Kubectl expose + should create services for rc [Conformance] + test/e2e/kubectl/kubectl.go:1413 +[BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:32:14.969 -Dec 14 16:32:14.969: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 16:32:14.972 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:32:15.006 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:32:15.011 -[It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:108 -STEP: Creating configMap with name projected-configmap-test-volume-map-2bc6a3b9-4f14-4d4c-86b1-c1a0fc1b8245 12/14/22 16:32:15.017 -STEP: Creating a pod to test consume configMaps 12/14/22 16:32:15.024 -Dec 14 16:32:15.039: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a" in namespace "projected-1674" to be "Succeeded or Failed" -Dec 14 16:32:15.044: INFO: Pod "pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.561362ms -Dec 14 16:32:17.061: INFO: Pod "pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021558134s -Dec 14 16:32:19.054: INFO: Pod "pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01501416s -STEP: Saw pod success 12/14/22 16:32:19.055 -Dec 14 16:32:19.055: INFO: Pod "pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a" satisfied condition "Succeeded or Failed" -Dec 14 16:32:19.061: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a container agnhost-container: -STEP: delete the pod 12/14/22 16:32:19.075 -Dec 14 16:32:19.122: INFO: Waiting for pod pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a to disappear -Dec 14 16:32:19.126: INFO: Pod pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a no longer exists -[AfterEach] [sig-storage] Projected configMap +STEP: Creating a kubernetes client 02/25/23 10:36:42.157 +Feb 25 10:36:42.158: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 10:36:42.169 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:36:42.228 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:36:42.241 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 +[It] should create services for rc [Conformance] + test/e2e/kubectl/kubectl.go:1413 +STEP: creating Agnhost RC 02/25/23 10:36:42.249 +Feb 25 10:36:42.251: INFO: namespace kubectl-7236 +Feb 25 10:36:42.251: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7236 create -f -' +Feb 25 10:36:44.386: INFO: stderr: "" +Feb 25 10:36:44.386: INFO: stdout: "replicationcontroller/agnhost-primary created\n" +STEP: Waiting for Agnhost primary to start. 02/25/23 10:36:44.386 +Feb 25 10:36:45.409: INFO: Selector matched 1 pods for map[app:agnhost] +Feb 25 10:36:45.409: INFO: Found 0 / 1 +Feb 25 10:36:46.400: INFO: Selector matched 1 pods for map[app:agnhost] +Feb 25 10:36:46.400: INFO: Found 1 / 1 +Feb 25 10:36:46.400: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 +Feb 25 10:36:46.408: INFO: Selector matched 1 pods for map[app:agnhost] +Feb 25 10:36:46.408: INFO: ForEach: Found 1 pods from the filter. Now looping through them. +Feb 25 10:36:46.408: INFO: wait on agnhost-primary startup in kubectl-7236 +Feb 25 10:36:46.409: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7236 logs agnhost-primary-sq95x agnhost-primary' +Feb 25 10:36:46.764: INFO: stderr: "" +Feb 25 10:36:46.764: INFO: stdout: "Paused\n" +STEP: exposing RC 02/25/23 10:36:46.764 +Feb 25 10:36:46.765: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7236 expose rc agnhost-primary --name=rm2 --port=1234 --target-port=6379' +Feb 25 10:36:46.978: INFO: stderr: "" +Feb 25 10:36:46.978: INFO: stdout: "service/rm2 exposed\n" +Feb 25 10:36:46.997: INFO: Service rm2 in namespace kubectl-7236 found. +STEP: exposing service 02/25/23 10:36:49.023 +Feb 25 10:36:49.023: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7236 expose service rm2 --name=rm3 --port=2345 --target-port=6379' +Feb 25 10:36:49.280: INFO: stderr: "" +Feb 25 10:36:49.280: INFO: stdout: "service/rm3 exposed\n" +Feb 25 10:36:49.309: INFO: Service rm3 in namespace kubectl-7236 found. +[AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 -Dec 14 16:32:19.126: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-1674" for this suite. 12/14/22 16:32:19.131 -{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]","completed":164,"skipped":2926,"failed":0} +Feb 25 10:36:51.369: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-7236" for this suite. 02/25/23 10:36:51.389 +{"msg":"PASSED [sig-cli] Kubectl client Kubectl expose should create services for rc [Conformance]","completed":162,"skipped":3007,"failed":0} ------------------------------ -• [4.196 seconds] -[sig-storage] Projected configMap -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:108 +• [SLOW TEST] [9.257 seconds] +[sig-cli] Kubectl client +test/e2e/kubectl/framework.go:23 + Kubectl expose + test/e2e/kubectl/kubectl.go:1407 + should create services for rc [Conformance] + test/e2e/kubectl/kubectl.go:1413 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected configMap + [BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:32:14.969 - Dec 14 16:32:14.969: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 16:32:14.972 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:32:15.006 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:32:15.011 - [It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:108 - STEP: Creating configMap with name projected-configmap-test-volume-map-2bc6a3b9-4f14-4d4c-86b1-c1a0fc1b8245 12/14/22 16:32:15.017 - STEP: Creating a pod to test consume configMaps 12/14/22 16:32:15.024 - Dec 14 16:32:15.039: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a" in namespace "projected-1674" to be "Succeeded or Failed" - Dec 14 16:32:15.044: INFO: Pod "pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.561362ms - Dec 14 16:32:17.061: INFO: Pod "pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021558134s - Dec 14 16:32:19.054: INFO: Pod "pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01501416s - STEP: Saw pod success 12/14/22 16:32:19.055 - Dec 14 16:32:19.055: INFO: Pod "pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a" satisfied condition "Succeeded or Failed" - Dec 14 16:32:19.061: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a container agnhost-container: - STEP: delete the pod 12/14/22 16:32:19.075 - Dec 14 16:32:19.122: INFO: Waiting for pod pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a to disappear - Dec 14 16:32:19.126: INFO: Pod pod-projected-configmaps-1556db8e-14a7-411f-a8b4-ff41ad5ef15a no longer exists - [AfterEach] [sig-storage] Projected configMap + STEP: Creating a kubernetes client 02/25/23 10:36:42.157 + Feb 25 10:36:42.158: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 10:36:42.169 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:36:42.228 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:36:42.241 + [BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 + [It] should create services for rc [Conformance] + test/e2e/kubectl/kubectl.go:1413 + STEP: creating Agnhost RC 02/25/23 10:36:42.249 + Feb 25 10:36:42.251: INFO: namespace kubectl-7236 + Feb 25 10:36:42.251: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7236 create -f -' + Feb 25 10:36:44.386: INFO: stderr: "" + Feb 25 10:36:44.386: INFO: stdout: "replicationcontroller/agnhost-primary created\n" + STEP: Waiting for Agnhost primary to start. 02/25/23 10:36:44.386 + Feb 25 10:36:45.409: INFO: Selector matched 1 pods for map[app:agnhost] + Feb 25 10:36:45.409: INFO: Found 0 / 1 + Feb 25 10:36:46.400: INFO: Selector matched 1 pods for map[app:agnhost] + Feb 25 10:36:46.400: INFO: Found 1 / 1 + Feb 25 10:36:46.400: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 + Feb 25 10:36:46.408: INFO: Selector matched 1 pods for map[app:agnhost] + Feb 25 10:36:46.408: INFO: ForEach: Found 1 pods from the filter. Now looping through them. + Feb 25 10:36:46.408: INFO: wait on agnhost-primary startup in kubectl-7236 + Feb 25 10:36:46.409: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7236 logs agnhost-primary-sq95x agnhost-primary' + Feb 25 10:36:46.764: INFO: stderr: "" + Feb 25 10:36:46.764: INFO: stdout: "Paused\n" + STEP: exposing RC 02/25/23 10:36:46.764 + Feb 25 10:36:46.765: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7236 expose rc agnhost-primary --name=rm2 --port=1234 --target-port=6379' + Feb 25 10:36:46.978: INFO: stderr: "" + Feb 25 10:36:46.978: INFO: stdout: "service/rm2 exposed\n" + Feb 25 10:36:46.997: INFO: Service rm2 in namespace kubectl-7236 found. + STEP: exposing service 02/25/23 10:36:49.023 + Feb 25 10:36:49.023: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-7236 expose service rm2 --name=rm3 --port=2345 --target-port=6379' + Feb 25 10:36:49.280: INFO: stderr: "" + Feb 25 10:36:49.280: INFO: stdout: "service/rm3 exposed\n" + Feb 25 10:36:49.309: INFO: Service rm3 in namespace kubectl-7236 found. + [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 16:32:19.126: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-1674" for this suite. 12/14/22 16:32:19.131 + Feb 25 10:36:51.369: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-7236" for this suite. 02/25/23 10:36:51.389 << End Captured GinkgoWriter Output ------------------------------ -[sig-network] Networking Granular Checks: Pods - should function for intra-pod communication: http [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:82 -[BeforeEach] [sig-network] Networking +SSSSSSSSS +------------------------------ +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + should mutate custom resource with different stored version [Conformance] + test/e2e/apimachinery/webhook.go:322 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:32:19.167 -Dec 14 16:32:19.167: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename pod-network-test 12/14/22 16:32:19.17 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:32:19.197 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:32:19.201 -[It] should function for intra-pod communication: http [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:82 -STEP: Performing setup for networking test in namespace pod-network-test-9281 12/14/22 16:32:19.205 -STEP: creating a selector 12/14/22 16:32:19.207 -STEP: Creating the service pods in kubernetes 12/14/22 16:32:19.207 -Dec 14 16:32:19.207: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable -Dec 14 16:32:19.261: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-9281" to be "running and ready" -Dec 14 16:32:19.268: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 6.741381ms -Dec 14 16:32:19.268: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:32:21.276: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 2.014310305s -Dec 14 16:32:21.276: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 16:32:23.276: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.014914032s -Dec 14 16:32:23.276: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 16:32:25.275: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.013720702s -Dec 14 16:32:25.275: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 16:32:27.279: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.01813781s -Dec 14 16:32:27.279: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 16:32:29.277: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.015314636s -Dec 14 16:32:29.277: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 16:32:31.279: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 12.017803439s -Dec 14 16:32:31.279: INFO: The phase of Pod netserver-0 is Running (Ready = true) -Dec 14 16:32:31.279: INFO: Pod "netserver-0" satisfied condition "running and ready" -Dec 14 16:32:31.287: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-9281" to be "running and ready" -Dec 14 16:32:31.292: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 4.891967ms -Dec 14 16:32:31.292: INFO: The phase of Pod netserver-1 is Running (Ready = true) -Dec 14 16:32:31.292: INFO: Pod "netserver-1" satisfied condition "running and ready" -Dec 14 16:32:31.296: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-9281" to be "running and ready" -Dec 14 16:32:31.306: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 9.891827ms -Dec 14 16:32:31.306: INFO: The phase of Pod netserver-2 is Running (Ready = false) -Dec 14 16:32:33.316: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 2.019837069s -Dec 14 16:32:33.316: INFO: The phase of Pod netserver-2 is Running (Ready = false) -Dec 14 16:32:35.317: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 4.020777202s -Dec 14 16:32:35.317: INFO: The phase of Pod netserver-2 is Running (Ready = false) -Dec 14 16:32:37.317: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 6.02089173s -Dec 14 16:32:37.317: INFO: The phase of Pod netserver-2 is Running (Ready = false) -Dec 14 16:32:39.313: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 8.017131998s -Dec 14 16:32:39.313: INFO: The phase of Pod netserver-2 is Running (Ready = false) -Dec 14 16:32:41.316: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 10.019880624s -Dec 14 16:32:41.316: INFO: The phase of Pod netserver-2 is Running (Ready = true) -Dec 14 16:32:41.316: INFO: Pod "netserver-2" satisfied condition "running and ready" -STEP: Creating test pods 12/14/22 16:32:41.321 -Dec 14 16:32:41.334: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-9281" to be "running" -Dec 14 16:32:41.341: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 7.394049ms -Dec 14 16:32:43.349: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01498729s -Dec 14 16:32:45.351: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.016654602s -Dec 14 16:32:45.351: INFO: Pod "test-container-pod" satisfied condition "running" -Dec 14 16:32:45.358: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 -Dec 14 16:32:45.358: INFO: Breadth first check of 10.233.64.105 on host 192.168.121.249... -Dec 14 16:32:45.363: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.247:9080/dial?request=hostname&protocol=http&host=10.233.64.105&port=8083&tries=1'] Namespace:pod-network-test-9281 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:32:45.364: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:32:45.366: INFO: ExecWithOptions: Clientset creation -Dec 14 16:32:45.367: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-9281/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.247%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dhttp%26host%3D10.233.64.105%26port%3D8083%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) -Dec 14 16:32:45.510: INFO: Waiting for responses: map[] -Dec 14 16:32:45.510: INFO: reached 10.233.64.105 after 0/1 tries -Dec 14 16:32:45.510: INFO: Breadth first check of 10.233.65.106 on host 192.168.121.197... -Dec 14 16:32:45.519: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.247:9080/dial?request=hostname&protocol=http&host=10.233.65.106&port=8083&tries=1'] Namespace:pod-network-test-9281 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:32:45.519: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:32:45.521: INFO: ExecWithOptions: Clientset creation -Dec 14 16:32:45.521: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-9281/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.247%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dhttp%26host%3D10.233.65.106%26port%3D8083%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) -Dec 14 16:32:45.610: INFO: Waiting for responses: map[] -Dec 14 16:32:45.610: INFO: reached 10.233.65.106 after 0/1 tries -Dec 14 16:32:45.610: INFO: Breadth first check of 10.233.66.246 on host 192.168.121.244... -Dec 14 16:32:45.616: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.247:9080/dial?request=hostname&protocol=http&host=10.233.66.246&port=8083&tries=1'] Namespace:pod-network-test-9281 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:32:45.616: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:32:45.619: INFO: ExecWithOptions: Clientset creation -Dec 14 16:32:45.619: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-9281/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.247%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dhttp%26host%3D10.233.66.246%26port%3D8083%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) -Dec 14 16:32:45.727: INFO: Waiting for responses: map[] -Dec 14 16:32:45.727: INFO: reached 10.233.66.246 after 0/1 tries -Dec 14 16:32:45.727: INFO: Going to retry 0 out of 3 pods.... -[AfterEach] [sig-network] Networking +STEP: Creating a kubernetes client 02/25/23 10:36:51.422 +Feb 25 10:36:51.422: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 10:36:51.425 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:36:51.481 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:36:51.504 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 +STEP: Setting up server cert 02/25/23 10:36:51.577 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:36:52.522 +STEP: Deploying the webhook pod 02/25/23 10:36:52.536 +STEP: Wait for the deployment to be ready 02/25/23 10:36:52.568 +Feb 25 10:36:52.595: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set +Feb 25 10:36:54.632: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 36, 52, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 36, 52, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 36, 52, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 36, 52, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} +STEP: Deploying the webhook service 02/25/23 10:36:56.647 +STEP: Verifying the service has paired with the endpoint 02/25/23 10:36:56.684 +Feb 25 10:36:57.685: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should mutate custom resource with different stored version [Conformance] + test/e2e/apimachinery/webhook.go:322 +Feb 25 10:36:57.707: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Registering the mutating webhook for custom resource e2e-test-webhook-6703-crds.webhook.example.com via the AdmissionRegistration API 02/25/23 10:36:58.28 +STEP: Creating a custom resource while v1 is storage version 02/25/23 10:36:58.362 +STEP: Patching Custom Resource Definition to set v2 as storage 02/25/23 10:37:00.808 +STEP: Patching the custom resource while v2 is storage version 02/25/23 10:37:00.846 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:32:45.728: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pod-network-test-9281" for this suite. 12/14/22 16:32:45.735 -{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [NodeConformance] [Conformance]","completed":165,"skipped":2926,"failed":0} +Feb 25 10:37:01.976: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-5839" for this suite. 02/25/23 10:37:01.99 +STEP: Destroying namespace "webhook-5839-markers" for this suite. 02/25/23 10:37:02.01 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with different stored version [Conformance]","completed":163,"skipped":3016,"failed":0} ------------------------------ -• [SLOW TEST] [26.579 seconds] -[sig-network] Networking -test/e2e/common/network/framework.go:23 - Granular Checks: Pods - test/e2e/common/network/networking.go:32 - should function for intra-pod communication: http [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:82 +• [SLOW TEST] [10.767 seconds] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + should mutate custom resource with different stored version [Conformance] + test/e2e/apimachinery/webhook.go:322 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Networking + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:32:19.167 - Dec 14 16:32:19.167: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename pod-network-test 12/14/22 16:32:19.17 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:32:19.197 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:32:19.201 - [It] should function for intra-pod communication: http [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:82 - STEP: Performing setup for networking test in namespace pod-network-test-9281 12/14/22 16:32:19.205 - STEP: creating a selector 12/14/22 16:32:19.207 - STEP: Creating the service pods in kubernetes 12/14/22 16:32:19.207 - Dec 14 16:32:19.207: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable - Dec 14 16:32:19.261: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-9281" to be "running and ready" - Dec 14 16:32:19.268: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 6.741381ms - Dec 14 16:32:19.268: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:32:21.276: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 2.014310305s - Dec 14 16:32:21.276: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 16:32:23.276: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.014914032s - Dec 14 16:32:23.276: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 16:32:25.275: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.013720702s - Dec 14 16:32:25.275: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 16:32:27.279: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.01813781s - Dec 14 16:32:27.279: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 16:32:29.277: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.015314636s - Dec 14 16:32:29.277: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 16:32:31.279: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 12.017803439s - Dec 14 16:32:31.279: INFO: The phase of Pod netserver-0 is Running (Ready = true) - Dec 14 16:32:31.279: INFO: Pod "netserver-0" satisfied condition "running and ready" - Dec 14 16:32:31.287: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-9281" to be "running and ready" - Dec 14 16:32:31.292: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 4.891967ms - Dec 14 16:32:31.292: INFO: The phase of Pod netserver-1 is Running (Ready = true) - Dec 14 16:32:31.292: INFO: Pod "netserver-1" satisfied condition "running and ready" - Dec 14 16:32:31.296: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-9281" to be "running and ready" - Dec 14 16:32:31.306: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 9.891827ms - Dec 14 16:32:31.306: INFO: The phase of Pod netserver-2 is Running (Ready = false) - Dec 14 16:32:33.316: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 2.019837069s - Dec 14 16:32:33.316: INFO: The phase of Pod netserver-2 is Running (Ready = false) - Dec 14 16:32:35.317: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 4.020777202s - Dec 14 16:32:35.317: INFO: The phase of Pod netserver-2 is Running (Ready = false) - Dec 14 16:32:37.317: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 6.02089173s - Dec 14 16:32:37.317: INFO: The phase of Pod netserver-2 is Running (Ready = false) - Dec 14 16:32:39.313: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 8.017131998s - Dec 14 16:32:39.313: INFO: The phase of Pod netserver-2 is Running (Ready = false) - Dec 14 16:32:41.316: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 10.019880624s - Dec 14 16:32:41.316: INFO: The phase of Pod netserver-2 is Running (Ready = true) - Dec 14 16:32:41.316: INFO: Pod "netserver-2" satisfied condition "running and ready" - STEP: Creating test pods 12/14/22 16:32:41.321 - Dec 14 16:32:41.334: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-9281" to be "running" - Dec 14 16:32:41.341: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 7.394049ms - Dec 14 16:32:43.349: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01498729s - Dec 14 16:32:45.351: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.016654602s - Dec 14 16:32:45.351: INFO: Pod "test-container-pod" satisfied condition "running" - Dec 14 16:32:45.358: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 - Dec 14 16:32:45.358: INFO: Breadth first check of 10.233.64.105 on host 192.168.121.249... - Dec 14 16:32:45.363: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.247:9080/dial?request=hostname&protocol=http&host=10.233.64.105&port=8083&tries=1'] Namespace:pod-network-test-9281 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:32:45.364: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:32:45.366: INFO: ExecWithOptions: Clientset creation - Dec 14 16:32:45.367: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-9281/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.247%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dhttp%26host%3D10.233.64.105%26port%3D8083%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) - Dec 14 16:32:45.510: INFO: Waiting for responses: map[] - Dec 14 16:32:45.510: INFO: reached 10.233.64.105 after 0/1 tries - Dec 14 16:32:45.510: INFO: Breadth first check of 10.233.65.106 on host 192.168.121.197... - Dec 14 16:32:45.519: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.247:9080/dial?request=hostname&protocol=http&host=10.233.65.106&port=8083&tries=1'] Namespace:pod-network-test-9281 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:32:45.519: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:32:45.521: INFO: ExecWithOptions: Clientset creation - Dec 14 16:32:45.521: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-9281/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.247%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dhttp%26host%3D10.233.65.106%26port%3D8083%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) - Dec 14 16:32:45.610: INFO: Waiting for responses: map[] - Dec 14 16:32:45.610: INFO: reached 10.233.65.106 after 0/1 tries - Dec 14 16:32:45.610: INFO: Breadth first check of 10.233.66.246 on host 192.168.121.244... - Dec 14 16:32:45.616: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.247:9080/dial?request=hostname&protocol=http&host=10.233.66.246&port=8083&tries=1'] Namespace:pod-network-test-9281 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:32:45.616: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:32:45.619: INFO: ExecWithOptions: Clientset creation - Dec 14 16:32:45.619: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-9281/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.247%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dhttp%26host%3D10.233.66.246%26port%3D8083%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) - Dec 14 16:32:45.727: INFO: Waiting for responses: map[] - Dec 14 16:32:45.727: INFO: reached 10.233.66.246 after 0/1 tries - Dec 14 16:32:45.727: INFO: Going to retry 0 out of 3 pods.... - [AfterEach] [sig-network] Networking + STEP: Creating a kubernetes client 02/25/23 10:36:51.422 + Feb 25 10:36:51.422: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 10:36:51.425 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:36:51.481 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:36:51.504 + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 + STEP: Setting up server cert 02/25/23 10:36:51.577 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:36:52.522 + STEP: Deploying the webhook pod 02/25/23 10:36:52.536 + STEP: Wait for the deployment to be ready 02/25/23 10:36:52.568 + Feb 25 10:36:52.595: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set + Feb 25 10:36:54.632: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 36, 52, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 36, 52, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 36, 52, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 36, 52, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} + STEP: Deploying the webhook service 02/25/23 10:36:56.647 + STEP: Verifying the service has paired with the endpoint 02/25/23 10:36:56.684 + Feb 25 10:36:57.685: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] should mutate custom resource with different stored version [Conformance] + test/e2e/apimachinery/webhook.go:322 + Feb 25 10:36:57.707: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Registering the mutating webhook for custom resource e2e-test-webhook-6703-crds.webhook.example.com via the AdmissionRegistration API 02/25/23 10:36:58.28 + STEP: Creating a custom resource while v1 is storage version 02/25/23 10:36:58.362 + STEP: Patching Custom Resource Definition to set v2 as storage 02/25/23 10:37:00.808 + STEP: Patching the custom resource while v2 is storage version 02/25/23 10:37:00.846 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:32:45.728: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "pod-network-test-9281" for this suite. 12/14/22 16:32:45.735 + Feb 25 10:37:01.976: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-5839" for this suite. 02/25/23 10:37:01.99 + STEP: Destroying namespace "webhook-5839-markers" for this suite. 02/25/23 10:37:02.01 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 << End Captured GinkgoWriter Output ------------------------------ -[sig-api-machinery] Watchers - should observe an object deletion if it stops meeting the requirements of the selector [Conformance] - test/e2e/apimachinery/watch.go:257 -[BeforeEach] [sig-api-machinery] Watchers +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-node] Containers + should be able to override the image's default command (container entrypoint) [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:72 +[BeforeEach] [sig-node] Containers test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:32:45.748 -Dec 14 16:32:45.748: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename watch 12/14/22 16:32:45.754 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:32:45.797 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:32:45.801 -[It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance] - test/e2e/apimachinery/watch.go:257 -STEP: creating a watch on configmaps with a certain label 12/14/22 16:32:45.804 -STEP: creating a new configmap 12/14/22 16:32:45.806 -STEP: modifying the configmap once 12/14/22 16:32:45.819 -STEP: changing the label value of the configmap 12/14/22 16:32:45.836 -STEP: Expecting to observe a delete notification for the watched object 12/14/22 16:32:45.845 -Dec 14 16:32:45.845: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-7087 ad9eb52b-2e0a-4871-a941-6204c94c7523 19013 0 2022-12-14 16:32:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2022-12-14 16:32:45 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} -Dec 14 16:32:45.846: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-7087 ad9eb52b-2e0a-4871-a941-6204c94c7523 19014 0 2022-12-14 16:32:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2022-12-14 16:32:45 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} -Dec 14 16:32:45.846: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-7087 ad9eb52b-2e0a-4871-a941-6204c94c7523 19015 0 2022-12-14 16:32:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2022-12-14 16:32:45 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} -STEP: modifying the configmap a second time 12/14/22 16:32:45.846 -STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements 12/14/22 16:32:45.855 -STEP: changing the label value of the configmap back 12/14/22 16:32:55.855 -STEP: modifying the configmap a third time 12/14/22 16:32:55.879 -STEP: deleting the configmap 12/14/22 16:32:55.889 -STEP: Expecting to observe an add notification for the watched object when the label value was restored 12/14/22 16:32:55.897 -Dec 14 16:32:55.897: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-7087 ad9eb52b-2e0a-4871-a941-6204c94c7523 19069 0 2022-12-14 16:32:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2022-12-14 16:32:55 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} -Dec 14 16:32:55.897: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-7087 ad9eb52b-2e0a-4871-a941-6204c94c7523 19070 0 2022-12-14 16:32:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2022-12-14 16:32:55 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},Immutable:nil,} -Dec 14 16:32:55.897: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-7087 ad9eb52b-2e0a-4871-a941-6204c94c7523 19071 0 2022-12-14 16:32:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2022-12-14 16:32:55 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},Immutable:nil,} -[AfterEach] [sig-api-machinery] Watchers +STEP: Creating a kubernetes client 02/25/23 10:37:02.22 +Feb 25 10:37:02.220: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename containers 02/25/23 10:37:02.231 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:37:02.298 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:37:02.306 +[It] should be able to override the image's default command (container entrypoint) [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:72 +STEP: Creating a pod to test override command 02/25/23 10:37:02.314 +Feb 25 10:37:02.362: INFO: Waiting up to 5m0s for pod "client-containers-c1d67302-1470-41ec-b942-9063ce800cf5" in namespace "containers-188" to be "Succeeded or Failed" +Feb 25 10:37:02.391: INFO: Pod "client-containers-c1d67302-1470-41ec-b942-9063ce800cf5": Phase="Pending", Reason="", readiness=false. Elapsed: 28.424229ms +Feb 25 10:37:04.412: INFO: Pod "client-containers-c1d67302-1470-41ec-b942-9063ce800cf5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.049401205s +Feb 25 10:37:06.398: INFO: Pod "client-containers-c1d67302-1470-41ec-b942-9063ce800cf5": Phase="Running", Reason="", readiness=false. Elapsed: 4.036206235s +Feb 25 10:37:08.404: INFO: Pod "client-containers-c1d67302-1470-41ec-b942-9063ce800cf5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.041851549s +STEP: Saw pod success 02/25/23 10:37:08.404 +Feb 25 10:37:08.405: INFO: Pod "client-containers-c1d67302-1470-41ec-b942-9063ce800cf5" satisfied condition "Succeeded or Failed" +Feb 25 10:37:08.416: INFO: Trying to get logs from node diek3dua9gay-3 pod client-containers-c1d67302-1470-41ec-b942-9063ce800cf5 container agnhost-container: +STEP: delete the pod 02/25/23 10:37:08.445 +Feb 25 10:37:08.478: INFO: Waiting for pod client-containers-c1d67302-1470-41ec-b942-9063ce800cf5 to disappear +Feb 25 10:37:08.486: INFO: Pod client-containers-c1d67302-1470-41ec-b942-9063ce800cf5 no longer exists +[AfterEach] [sig-node] Containers test/e2e/framework/framework.go:187 -Dec 14 16:32:55.898: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "watch-7087" for this suite. 12/14/22 16:32:55.903 -{"msg":"PASSED [sig-api-machinery] Watchers should observe an object deletion if it stops meeting the requirements of the selector [Conformance]","completed":166,"skipped":2926,"failed":0} +Feb 25 10:37:08.487: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "containers-188" for this suite. 02/25/23 10:37:08.499 +{"msg":"PASSED [sig-node] Containers should be able to override the image's default command (container entrypoint) [NodeConformance] [Conformance]","completed":164,"skipped":3058,"failed":0} ------------------------------ -• [SLOW TEST] [10.163 seconds] -[sig-api-machinery] Watchers -test/e2e/apimachinery/framework.go:23 - should observe an object deletion if it stops meeting the requirements of the selector [Conformance] - test/e2e/apimachinery/watch.go:257 +• [SLOW TEST] [6.295 seconds] +[sig-node] Containers +test/e2e/common/node/framework.go:23 + should be able to override the image's default command (container entrypoint) [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:72 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Watchers + [BeforeEach] [sig-node] Containers test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:32:45.748 - Dec 14 16:32:45.748: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename watch 12/14/22 16:32:45.754 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:32:45.797 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:32:45.801 - [It] should observe an object deletion if it stops meeting the requirements of the selector [Conformance] - test/e2e/apimachinery/watch.go:257 - STEP: creating a watch on configmaps with a certain label 12/14/22 16:32:45.804 - STEP: creating a new configmap 12/14/22 16:32:45.806 - STEP: modifying the configmap once 12/14/22 16:32:45.819 - STEP: changing the label value of the configmap 12/14/22 16:32:45.836 - STEP: Expecting to observe a delete notification for the watched object 12/14/22 16:32:45.845 - Dec 14 16:32:45.845: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-7087 ad9eb52b-2e0a-4871-a941-6204c94c7523 19013 0 2022-12-14 16:32:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2022-12-14 16:32:45 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} - Dec 14 16:32:45.846: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-7087 ad9eb52b-2e0a-4871-a941-6204c94c7523 19014 0 2022-12-14 16:32:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2022-12-14 16:32:45 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} - Dec 14 16:32:45.846: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-7087 ad9eb52b-2e0a-4871-a941-6204c94c7523 19015 0 2022-12-14 16:32:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2022-12-14 16:32:45 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} - STEP: modifying the configmap a second time 12/14/22 16:32:45.846 - STEP: Expecting not to observe a notification because the object no longer meets the selector's requirements 12/14/22 16:32:45.855 - STEP: changing the label value of the configmap back 12/14/22 16:32:55.855 - STEP: modifying the configmap a third time 12/14/22 16:32:55.879 - STEP: deleting the configmap 12/14/22 16:32:55.889 - STEP: Expecting to observe an add notification for the watched object when the label value was restored 12/14/22 16:32:55.897 - Dec 14 16:32:55.897: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-7087 ad9eb52b-2e0a-4871-a941-6204c94c7523 19069 0 2022-12-14 16:32:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2022-12-14 16:32:55 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} - Dec 14 16:32:55.897: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-7087 ad9eb52b-2e0a-4871-a941-6204c94c7523 19070 0 2022-12-14 16:32:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2022-12-14 16:32:55 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},Immutable:nil,} - Dec 14 16:32:55.897: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-label-changed watch-7087 ad9eb52b-2e0a-4871-a941-6204c94c7523 19071 0 2022-12-14 16:32:45 +0000 UTC map[watch-this-configmap:label-changed-and-restored] map[] [] [] [{e2e.test Update v1 2022-12-14 16:32:55 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 3,},BinaryData:map[string][]byte{},Immutable:nil,} - [AfterEach] [sig-api-machinery] Watchers + STEP: Creating a kubernetes client 02/25/23 10:37:02.22 + Feb 25 10:37:02.220: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename containers 02/25/23 10:37:02.231 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:37:02.298 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:37:02.306 + [It] should be able to override the image's default command (container entrypoint) [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:72 + STEP: Creating a pod to test override command 02/25/23 10:37:02.314 + Feb 25 10:37:02.362: INFO: Waiting up to 5m0s for pod "client-containers-c1d67302-1470-41ec-b942-9063ce800cf5" in namespace "containers-188" to be "Succeeded or Failed" + Feb 25 10:37:02.391: INFO: Pod "client-containers-c1d67302-1470-41ec-b942-9063ce800cf5": Phase="Pending", Reason="", readiness=false. Elapsed: 28.424229ms + Feb 25 10:37:04.412: INFO: Pod "client-containers-c1d67302-1470-41ec-b942-9063ce800cf5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.049401205s + Feb 25 10:37:06.398: INFO: Pod "client-containers-c1d67302-1470-41ec-b942-9063ce800cf5": Phase="Running", Reason="", readiness=false. Elapsed: 4.036206235s + Feb 25 10:37:08.404: INFO: Pod "client-containers-c1d67302-1470-41ec-b942-9063ce800cf5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.041851549s + STEP: Saw pod success 02/25/23 10:37:08.404 + Feb 25 10:37:08.405: INFO: Pod "client-containers-c1d67302-1470-41ec-b942-9063ce800cf5" satisfied condition "Succeeded or Failed" + Feb 25 10:37:08.416: INFO: Trying to get logs from node diek3dua9gay-3 pod client-containers-c1d67302-1470-41ec-b942-9063ce800cf5 container agnhost-container: + STEP: delete the pod 02/25/23 10:37:08.445 + Feb 25 10:37:08.478: INFO: Waiting for pod client-containers-c1d67302-1470-41ec-b942-9063ce800cf5 to disappear + Feb 25 10:37:08.486: INFO: Pod client-containers-c1d67302-1470-41ec-b942-9063ce800cf5 no longer exists + [AfterEach] [sig-node] Containers test/e2e/framework/framework.go:187 - Dec 14 16:32:55.898: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "watch-7087" for this suite. 12/14/22 16:32:55.903 + Feb 25 10:37:08.487: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "containers-188" for this suite. 02/25/23 10:37:08.499 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-auth] Certificates API [Privileged:ClusterAdmin] - should support CSR API operations [Conformance] - test/e2e/auth/certificates.go:200 -[BeforeEach] [sig-auth] Certificates API [Privileged:ClusterAdmin] +[sig-api-machinery] Garbage collector + should delete pods created by rc when not orphaning [Conformance] + test/e2e/apimachinery/garbage_collector.go:312 +[BeforeEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:32:55.931 -Dec 14 16:32:55.932: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename certificates 12/14/22 16:32:55.933 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:32:55.958 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:32:55.961 -[It] should support CSR API operations [Conformance] - test/e2e/auth/certificates.go:200 -STEP: getting /apis 12/14/22 16:32:57.716 -STEP: getting /apis/certificates.k8s.io 12/14/22 16:32:57.723 -STEP: getting /apis/certificates.k8s.io/v1 12/14/22 16:32:57.725 -STEP: creating 12/14/22 16:32:57.727 -STEP: getting 12/14/22 16:32:57.783 -STEP: listing 12/14/22 16:32:57.792 -STEP: watching 12/14/22 16:32:57.8 -Dec 14 16:32:57.800: INFO: starting watch -STEP: patching 12/14/22 16:32:57.802 -STEP: updating 12/14/22 16:32:57.813 -Dec 14 16:32:57.827: INFO: waiting for watch events with expected annotations -Dec 14 16:32:57.827: INFO: saw patched and updated annotations -STEP: getting /approval 12/14/22 16:32:57.827 -STEP: patching /approval 12/14/22 16:32:57.833 -STEP: updating /approval 12/14/22 16:32:57.844 -STEP: getting /status 12/14/22 16:32:57.852 -STEP: patching /status 12/14/22 16:32:57.857 -STEP: updating /status 12/14/22 16:32:57.888 -STEP: deleting 12/14/22 16:32:57.895 -STEP: deleting a collection 12/14/22 16:32:57.909 -[AfterEach] [sig-auth] Certificates API [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 10:37:08.527 +Feb 25 10:37:08.528: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename gc 02/25/23 10:37:08.533 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:37:08.563 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:37:08.571 +[It] should delete pods created by rc when not orphaning [Conformance] + test/e2e/apimachinery/garbage_collector.go:312 +STEP: create the rc 02/25/23 10:37:08.581 +STEP: delete the rc 02/25/23 10:37:13.622 +STEP: wait for all pods to be garbage collected 02/25/23 10:37:13.637 +STEP: Gathering metrics 02/25/23 10:37:18.659 +Feb 25 10:37:18.715: INFO: Waiting up to 5m0s for pod "kube-controller-manager-diek3dua9gay-2" in namespace "kube-system" to be "running and ready" +Feb 25 10:37:18.725: INFO: Pod "kube-controller-manager-diek3dua9gay-2": Phase="Running", Reason="", readiness=true. Elapsed: 9.994952ms +Feb 25 10:37:18.726: INFO: The phase of Pod kube-controller-manager-diek3dua9gay-2 is Running (Ready = true) +Feb 25 10:37:18.726: INFO: Pod "kube-controller-manager-diek3dua9gay-2" satisfied condition "running and ready" +Feb 25 10:37:18.850: INFO: For apiserver_request_total: +For apiserver_request_latency_seconds: +For apiserver_init_events_total: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: + +[AfterEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:187 -Dec 14 16:32:57.934: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "certificates-8504" for this suite. 12/14/22 16:32:57.94 -{"msg":"PASSED [sig-auth] Certificates API [Privileged:ClusterAdmin] should support CSR API operations [Conformance]","completed":167,"skipped":3002,"failed":0} +Feb 25 10:37:18.850: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "gc-9721" for this suite. 02/25/23 10:37:18.865 +{"msg":"PASSED [sig-api-machinery] Garbage collector should delete pods created by rc when not orphaning [Conformance]","completed":165,"skipped":3083,"failed":0} ------------------------------ -• [2.016 seconds] -[sig-auth] Certificates API [Privileged:ClusterAdmin] -test/e2e/auth/framework.go:23 - should support CSR API operations [Conformance] - test/e2e/auth/certificates.go:200 +• [SLOW TEST] [10.352 seconds] +[sig-api-machinery] Garbage collector +test/e2e/apimachinery/framework.go:23 + should delete pods created by rc when not orphaning [Conformance] + test/e2e/apimachinery/garbage_collector.go:312 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-auth] Certificates API [Privileged:ClusterAdmin] + [BeforeEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:32:55.931 - Dec 14 16:32:55.932: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename certificates 12/14/22 16:32:55.933 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:32:55.958 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:32:55.961 - [It] should support CSR API operations [Conformance] - test/e2e/auth/certificates.go:200 - STEP: getting /apis 12/14/22 16:32:57.716 - STEP: getting /apis/certificates.k8s.io 12/14/22 16:32:57.723 - STEP: getting /apis/certificates.k8s.io/v1 12/14/22 16:32:57.725 - STEP: creating 12/14/22 16:32:57.727 - STEP: getting 12/14/22 16:32:57.783 - STEP: listing 12/14/22 16:32:57.792 - STEP: watching 12/14/22 16:32:57.8 - Dec 14 16:32:57.800: INFO: starting watch - STEP: patching 12/14/22 16:32:57.802 - STEP: updating 12/14/22 16:32:57.813 - Dec 14 16:32:57.827: INFO: waiting for watch events with expected annotations - Dec 14 16:32:57.827: INFO: saw patched and updated annotations - STEP: getting /approval 12/14/22 16:32:57.827 - STEP: patching /approval 12/14/22 16:32:57.833 - STEP: updating /approval 12/14/22 16:32:57.844 - STEP: getting /status 12/14/22 16:32:57.852 - STEP: patching /status 12/14/22 16:32:57.857 - STEP: updating /status 12/14/22 16:32:57.888 - STEP: deleting 12/14/22 16:32:57.895 - STEP: deleting a collection 12/14/22 16:32:57.909 - [AfterEach] [sig-auth] Certificates API [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 10:37:08.527 + Feb 25 10:37:08.528: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename gc 02/25/23 10:37:08.533 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:37:08.563 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:37:08.571 + [It] should delete pods created by rc when not orphaning [Conformance] + test/e2e/apimachinery/garbage_collector.go:312 + STEP: create the rc 02/25/23 10:37:08.581 + STEP: delete the rc 02/25/23 10:37:13.622 + STEP: wait for all pods to be garbage collected 02/25/23 10:37:13.637 + STEP: Gathering metrics 02/25/23 10:37:18.659 + Feb 25 10:37:18.715: INFO: Waiting up to 5m0s for pod "kube-controller-manager-diek3dua9gay-2" in namespace "kube-system" to be "running and ready" + Feb 25 10:37:18.725: INFO: Pod "kube-controller-manager-diek3dua9gay-2": Phase="Running", Reason="", readiness=true. Elapsed: 9.994952ms + Feb 25 10:37:18.726: INFO: The phase of Pod kube-controller-manager-diek3dua9gay-2 is Running (Ready = true) + Feb 25 10:37:18.726: INFO: Pod "kube-controller-manager-diek3dua9gay-2" satisfied condition "running and ready" + Feb 25 10:37:18.850: INFO: For apiserver_request_total: + For apiserver_request_latency_seconds: + For apiserver_init_events_total: + For garbage_collector_attempt_to_delete_queue_latency: + For garbage_collector_attempt_to_delete_work_duration: + For garbage_collector_attempt_to_orphan_queue_latency: + For garbage_collector_attempt_to_orphan_work_duration: + For garbage_collector_dirty_processing_latency_microseconds: + For garbage_collector_event_processing_latency_microseconds: + For garbage_collector_graph_changes_queue_latency: + For garbage_collector_graph_changes_work_duration: + For garbage_collector_orphan_processing_latency_microseconds: + For namespace_queue_latency: + For namespace_queue_latency_sum: + For namespace_queue_latency_count: + For namespace_retries: + For namespace_work_duration: + For namespace_work_duration_sum: + For namespace_work_duration_count: + For function_duration_seconds: + For errors_total: + For evicted_pods_total: + + [AfterEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:187 - Dec 14 16:32:57.934: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "certificates-8504" for this suite. 12/14/22 16:32:57.94 + Feb 25 10:37:18.850: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "gc-9721" for this suite. 02/25/23 10:37:18.865 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSS +SSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] KubeletManagedEtcHosts - should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/kubelet_etc_hosts.go:63 -[BeforeEach] [sig-node] KubeletManagedEtcHosts +[sig-network] Services + should test the lifecycle of an Endpoint [Conformance] + test/e2e/network/service.go:3231 +[BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:32:57.951 -Dec 14 16:32:57.951: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts 12/14/22 16:32:57.952 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:32:57.967 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:32:57.969 -[It] should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/kubelet_etc_hosts.go:63 -STEP: Setting up the test 12/14/22 16:32:57.972 -STEP: Creating hostNetwork=false pod 12/14/22 16:32:57.972 -Dec 14 16:32:57.986: INFO: Waiting up to 5m0s for pod "test-pod" in namespace "e2e-kubelet-etc-hosts-2988" to be "running and ready" -Dec 14 16:32:57.991: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 4.633563ms -Dec 14 16:32:57.991: INFO: The phase of Pod test-pod is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:33:00.001: INFO: Pod "test-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.014730698s -Dec 14 16:33:00.001: INFO: The phase of Pod test-pod is Running (Ready = true) -Dec 14 16:33:00.001: INFO: Pod "test-pod" satisfied condition "running and ready" -STEP: Creating hostNetwork=true pod 12/14/22 16:33:00.01 -Dec 14 16:33:00.021: INFO: Waiting up to 5m0s for pod "test-host-network-pod" in namespace "e2e-kubelet-etc-hosts-2988" to be "running and ready" -Dec 14 16:33:00.026: INFO: Pod "test-host-network-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 4.781258ms -Dec 14 16:33:00.026: INFO: The phase of Pod test-host-network-pod is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:33:02.042: INFO: Pod "test-host-network-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.020044551s -Dec 14 16:33:02.042: INFO: The phase of Pod test-host-network-pod is Running (Ready = true) -Dec 14 16:33:02.042: INFO: Pod "test-host-network-pod" satisfied condition "running and ready" -STEP: Running the test 12/14/22 16:33:02.053 -STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false 12/14/22 16:33:02.053 -Dec 14 16:33:02.054: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:33:02.054: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:33:02.055: INFO: ExecWithOptions: Clientset creation -Dec 14 16:33:02.059: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-1&container=busybox-1&stderr=true&stdout=true) -Dec 14 16:33:02.210: INFO: Exec stderr: "" -Dec 14 16:33:02.210: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:33:02.210: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:33:02.212: INFO: ExecWithOptions: Clientset creation -Dec 14 16:33:02.212: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-1&container=busybox-1&stderr=true&stdout=true) -Dec 14 16:33:02.310: INFO: Exec stderr: "" -Dec 14 16:33:02.310: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:33:02.310: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:33:02.311: INFO: ExecWithOptions: Clientset creation -Dec 14 16:33:02.311: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-2&container=busybox-2&stderr=true&stdout=true) -Dec 14 16:33:02.426: INFO: Exec stderr: "" -Dec 14 16:33:02.426: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:33:02.426: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:33:02.427: INFO: ExecWithOptions: Clientset creation -Dec 14 16:33:02.427: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-2&container=busybox-2&stderr=true&stdout=true) -Dec 14 16:33:02.515: INFO: Exec stderr: "" -STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount 12/14/22 16:33:02.515 -Dec 14 16:33:02.515: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:33:02.515: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:33:02.516: INFO: ExecWithOptions: Clientset creation -Dec 14 16:33:02.516: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-3&container=busybox-3&stderr=true&stdout=true) -Dec 14 16:33:02.656: INFO: Exec stderr: "" -Dec 14 16:33:02.656: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:33:02.656: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:33:02.658: INFO: ExecWithOptions: Clientset creation -Dec 14 16:33:02.658: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-3&container=busybox-3&stderr=true&stdout=true) -Dec 14 16:33:02.754: INFO: Exec stderr: "" -STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true 12/14/22 16:33:02.755 -Dec 14 16:33:02.755: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:33:02.755: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:33:02.756: INFO: ExecWithOptions: Clientset creation -Dec 14 16:33:02.756: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-1&container=busybox-1&stderr=true&stdout=true) -Dec 14 16:33:02.847: INFO: Exec stderr: "" -Dec 14 16:33:02.847: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:33:02.847: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:33:02.848: INFO: ExecWithOptions: Clientset creation -Dec 14 16:33:02.849: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-1&container=busybox-1&stderr=true&stdout=true) -Dec 14 16:33:02.941: INFO: Exec stderr: "" -Dec 14 16:33:02.941: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:33:02.941: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:33:02.943: INFO: ExecWithOptions: Clientset creation -Dec 14 16:33:02.944: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-2&container=busybox-2&stderr=true&stdout=true) -Dec 14 16:33:03.045: INFO: Exec stderr: "" -Dec 14 16:33:03.046: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:33:03.046: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:33:03.048: INFO: ExecWithOptions: Clientset creation -Dec 14 16:33:03.048: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-2&container=busybox-2&stderr=true&stdout=true) -Dec 14 16:33:03.128: INFO: Exec stderr: "" -[AfterEach] [sig-node] KubeletManagedEtcHosts +STEP: Creating a kubernetes client 02/25/23 10:37:18.883 +Feb 25 10:37:18.883: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 10:37:18.884 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:37:18.983 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:37:19.003 +[BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 +[It] should test the lifecycle of an Endpoint [Conformance] + test/e2e/network/service.go:3231 +STEP: creating an Endpoint 02/25/23 10:37:19.022 +STEP: waiting for available Endpoint 02/25/23 10:37:19.038 +STEP: listing all Endpoints 02/25/23 10:37:19.05 +STEP: updating the Endpoint 02/25/23 10:37:19.064 +STEP: fetching the Endpoint 02/25/23 10:37:19.089 +STEP: patching the Endpoint 02/25/23 10:37:19.1 +STEP: fetching the Endpoint 02/25/23 10:37:19.13 +STEP: deleting the Endpoint by Collection 02/25/23 10:37:19.139 +STEP: waiting for Endpoint deletion 02/25/23 10:37:19.161 +STEP: fetching the Endpoint 02/25/23 10:37:19.17 +[AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 16:33:03.129: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "e2e-kubelet-etc-hosts-2988" for this suite. 12/14/22 16:33:03.135 -{"msg":"PASSED [sig-node] KubeletManagedEtcHosts should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance]","completed":168,"skipped":3011,"failed":0} +Feb 25 10:37:19.184: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-9895" for this suite. 02/25/23 10:37:19.196 +[AfterEach] [sig-network] Services + test/e2e/network/service.go:762 +{"msg":"PASSED [sig-network] Services should test the lifecycle of an Endpoint [Conformance]","completed":166,"skipped":3102,"failed":0} ------------------------------ -• [SLOW TEST] [5.191 seconds] -[sig-node] KubeletManagedEtcHosts -test/e2e/common/node/framework.go:23 - should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/kubelet_etc_hosts.go:63 +• [0.334 seconds] +[sig-network] Services +test/e2e/network/common/framework.go:23 + should test the lifecycle of an Endpoint [Conformance] + test/e2e/network/service.go:3231 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] KubeletManagedEtcHosts + [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:32:57.951 - Dec 14 16:32:57.951: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename e2e-kubelet-etc-hosts 12/14/22 16:32:57.952 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:32:57.967 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:32:57.969 - [It] should test kubelet managed /etc/hosts file [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/kubelet_etc_hosts.go:63 - STEP: Setting up the test 12/14/22 16:32:57.972 - STEP: Creating hostNetwork=false pod 12/14/22 16:32:57.972 - Dec 14 16:32:57.986: INFO: Waiting up to 5m0s for pod "test-pod" in namespace "e2e-kubelet-etc-hosts-2988" to be "running and ready" - Dec 14 16:32:57.991: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 4.633563ms - Dec 14 16:32:57.991: INFO: The phase of Pod test-pod is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:33:00.001: INFO: Pod "test-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.014730698s - Dec 14 16:33:00.001: INFO: The phase of Pod test-pod is Running (Ready = true) - Dec 14 16:33:00.001: INFO: Pod "test-pod" satisfied condition "running and ready" - STEP: Creating hostNetwork=true pod 12/14/22 16:33:00.01 - Dec 14 16:33:00.021: INFO: Waiting up to 5m0s for pod "test-host-network-pod" in namespace "e2e-kubelet-etc-hosts-2988" to be "running and ready" - Dec 14 16:33:00.026: INFO: Pod "test-host-network-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 4.781258ms - Dec 14 16:33:00.026: INFO: The phase of Pod test-host-network-pod is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:33:02.042: INFO: Pod "test-host-network-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.020044551s - Dec 14 16:33:02.042: INFO: The phase of Pod test-host-network-pod is Running (Ready = true) - Dec 14 16:33:02.042: INFO: Pod "test-host-network-pod" satisfied condition "running and ready" - STEP: Running the test 12/14/22 16:33:02.053 - STEP: Verifying /etc/hosts of container is kubelet-managed for pod with hostNetwork=false 12/14/22 16:33:02.053 - Dec 14 16:33:02.054: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:33:02.054: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:33:02.055: INFO: ExecWithOptions: Clientset creation - Dec 14 16:33:02.059: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-1&container=busybox-1&stderr=true&stdout=true) - Dec 14 16:33:02.210: INFO: Exec stderr: "" - Dec 14 16:33:02.210: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:33:02.210: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:33:02.212: INFO: ExecWithOptions: Clientset creation - Dec 14 16:33:02.212: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-1&container=busybox-1&stderr=true&stdout=true) - Dec 14 16:33:02.310: INFO: Exec stderr: "" - Dec 14 16:33:02.310: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:33:02.310: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:33:02.311: INFO: ExecWithOptions: Clientset creation - Dec 14 16:33:02.311: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-2&container=busybox-2&stderr=true&stdout=true) - Dec 14 16:33:02.426: INFO: Exec stderr: "" - Dec 14 16:33:02.426: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:33:02.426: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:33:02.427: INFO: ExecWithOptions: Clientset creation - Dec 14 16:33:02.427: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-2&container=busybox-2&stderr=true&stdout=true) - Dec 14 16:33:02.515: INFO: Exec stderr: "" - STEP: Verifying /etc/hosts of container is not kubelet-managed since container specifies /etc/hosts mount 12/14/22 16:33:02.515 - Dec 14 16:33:02.515: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:33:02.515: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:33:02.516: INFO: ExecWithOptions: Clientset creation - Dec 14 16:33:02.516: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-3&container=busybox-3&stderr=true&stdout=true) - Dec 14 16:33:02.656: INFO: Exec stderr: "" - Dec 14 16:33:02.656: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-pod ContainerName:busybox-3 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:33:02.656: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:33:02.658: INFO: ExecWithOptions: Clientset creation - Dec 14 16:33:02.658: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-3&container=busybox-3&stderr=true&stdout=true) - Dec 14 16:33:02.754: INFO: Exec stderr: "" - STEP: Verifying /etc/hosts content of container is not kubelet-managed for pod with hostNetwork=true 12/14/22 16:33:02.755 - Dec 14 16:33:02.755: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:33:02.755: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:33:02.756: INFO: ExecWithOptions: Clientset creation - Dec 14 16:33:02.756: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-1&container=busybox-1&stderr=true&stdout=true) - Dec 14 16:33:02.847: INFO: Exec stderr: "" - Dec 14 16:33:02.847: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-host-network-pod ContainerName:busybox-1 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:33:02.847: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:33:02.848: INFO: ExecWithOptions: Clientset creation - Dec 14 16:33:02.849: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-1&container=busybox-1&stderr=true&stdout=true) - Dec 14 16:33:02.941: INFO: Exec stderr: "" - Dec 14 16:33:02.941: INFO: ExecWithOptions {Command:[cat /etc/hosts] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:33:02.941: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:33:02.943: INFO: ExecWithOptions: Clientset creation - Dec 14 16:33:02.944: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts&container=busybox-2&container=busybox-2&stderr=true&stdout=true) - Dec 14 16:33:03.045: INFO: Exec stderr: "" - Dec 14 16:33:03.046: INFO: ExecWithOptions {Command:[cat /etc/hosts-original] Namespace:e2e-kubelet-etc-hosts-2988 PodName:test-host-network-pod ContainerName:busybox-2 Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:33:03.046: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:33:03.048: INFO: ExecWithOptions: Clientset creation - Dec 14 16:33:03.048: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/e2e-kubelet-etc-hosts-2988/pods/test-host-network-pod/exec?command=cat&command=%2Fetc%2Fhosts-original&container=busybox-2&container=busybox-2&stderr=true&stdout=true) - Dec 14 16:33:03.128: INFO: Exec stderr: "" - [AfterEach] [sig-node] KubeletManagedEtcHosts + STEP: Creating a kubernetes client 02/25/23 10:37:18.883 + Feb 25 10:37:18.883: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 10:37:18.884 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:37:18.983 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:37:19.003 + [BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 + [It] should test the lifecycle of an Endpoint [Conformance] + test/e2e/network/service.go:3231 + STEP: creating an Endpoint 02/25/23 10:37:19.022 + STEP: waiting for available Endpoint 02/25/23 10:37:19.038 + STEP: listing all Endpoints 02/25/23 10:37:19.05 + STEP: updating the Endpoint 02/25/23 10:37:19.064 + STEP: fetching the Endpoint 02/25/23 10:37:19.089 + STEP: patching the Endpoint 02/25/23 10:37:19.1 + STEP: fetching the Endpoint 02/25/23 10:37:19.13 + STEP: deleting the Endpoint by Collection 02/25/23 10:37:19.139 + STEP: waiting for Endpoint deletion 02/25/23 10:37:19.161 + STEP: fetching the Endpoint 02/25/23 10:37:19.17 + [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 16:33:03.129: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "e2e-kubelet-etc-hosts-2988" for this suite. 12/14/22 16:33:03.135 + Feb 25 10:37:19.184: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-9895" for this suite. 02/25/23 10:37:19.196 + [AfterEach] [sig-network] Services + test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Secrets - should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:124 -[BeforeEach] [sig-storage] Secrets +[sig-apps] Job + should adopt matching orphans and release non-matching pods [Conformance] + test/e2e/apps/job.go:335 +[BeforeEach] [sig-apps] Job test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:33:03.15 -Dec 14 16:33:03.150: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename secrets 12/14/22 16:33:03.152 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:33:03.19 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:33:03.193 -[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:124 -STEP: Creating secret with name secret-test-297c87ac-6d23-49ac-b8ad-6670203fc668 12/14/22 16:33:03.198 -STEP: Creating a pod to test consume secrets 12/14/22 16:33:03.204 -Dec 14 16:33:03.218: INFO: Waiting up to 5m0s for pod "pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f" in namespace "secrets-4640" to be "Succeeded or Failed" -Dec 14 16:33:03.225: INFO: Pod "pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f": Phase="Pending", Reason="", readiness=false. Elapsed: 7.140027ms -Dec 14 16:33:05.231: INFO: Pod "pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013315786s -Dec 14 16:33:07.233: INFO: Pod "pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014845711s -STEP: Saw pod success 12/14/22 16:33:07.233 -Dec 14 16:33:07.233: INFO: Pod "pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f" satisfied condition "Succeeded or Failed" -Dec 14 16:33:07.238: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f container secret-volume-test: -STEP: delete the pod 12/14/22 16:33:07.249 -Dec 14 16:33:07.270: INFO: Waiting for pod pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f to disappear -Dec 14 16:33:07.279: INFO: Pod pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f no longer exists -[AfterEach] [sig-storage] Secrets +STEP: Creating a kubernetes client 02/25/23 10:37:19.241 +Feb 25 10:37:19.242: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename job 02/25/23 10:37:19.244 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:37:19.287 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:37:19.298 +[It] should adopt matching orphans and release non-matching pods [Conformance] + test/e2e/apps/job.go:335 +STEP: Creating a job 02/25/23 10:37:19.311 +STEP: Ensuring active pods == parallelism 02/25/23 10:37:19.339 +STEP: Orphaning one of the Job's Pods 02/25/23 10:37:23.348 +Feb 25 10:37:23.882: INFO: Successfully updated pod "adopt-release-5nfkf" +STEP: Checking that the Job readopts the Pod 02/25/23 10:37:23.883 +Feb 25 10:37:23.883: INFO: Waiting up to 15m0s for pod "adopt-release-5nfkf" in namespace "job-3771" to be "adopted" +Feb 25 10:37:23.904: INFO: Pod "adopt-release-5nfkf": Phase="Running", Reason="", readiness=true. Elapsed: 20.260094ms +Feb 25 10:37:25.916: INFO: Pod "adopt-release-5nfkf": Phase="Running", Reason="", readiness=true. Elapsed: 2.032376609s +Feb 25 10:37:25.916: INFO: Pod "adopt-release-5nfkf" satisfied condition "adopted" +STEP: Removing the labels from the Job's Pod 02/25/23 10:37:25.916 +Feb 25 10:37:26.447: INFO: Successfully updated pod "adopt-release-5nfkf" +STEP: Checking that the Job releases the Pod 02/25/23 10:37:26.447 +Feb 25 10:37:26.447: INFO: Waiting up to 15m0s for pod "adopt-release-5nfkf" in namespace "job-3771" to be "released" +Feb 25 10:37:26.465: INFO: Pod "adopt-release-5nfkf": Phase="Running", Reason="", readiness=true. Elapsed: 18.024647ms +Feb 25 10:37:28.477: INFO: Pod "adopt-release-5nfkf": Phase="Running", Reason="", readiness=true. Elapsed: 2.030329528s +Feb 25 10:37:28.477: INFO: Pod "adopt-release-5nfkf" satisfied condition "released" +[AfterEach] [sig-apps] Job test/e2e/framework/framework.go:187 -Dec 14 16:33:07.280: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-4640" for this suite. 12/14/22 16:33:07.285 -{"msg":"PASSED [sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","completed":169,"skipped":3056,"failed":0} +Feb 25 10:37:28.478: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "job-3771" for this suite. 02/25/23 10:37:28.493 +{"msg":"PASSED [sig-apps] Job should adopt matching orphans and release non-matching pods [Conformance]","completed":167,"skipped":3146,"failed":0} ------------------------------ -• [4.154 seconds] -[sig-storage] Secrets -test/e2e/common/storage/framework.go:23 - should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:124 +• [SLOW TEST] [9.296 seconds] +[sig-apps] Job +test/e2e/apps/framework.go:23 + should adopt matching orphans and release non-matching pods [Conformance] + test/e2e/apps/job.go:335 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Secrets + [BeforeEach] [sig-apps] Job test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:33:03.15 - Dec 14 16:33:03.150: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename secrets 12/14/22 16:33:03.152 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:33:03.19 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:33:03.193 - [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:124 - STEP: Creating secret with name secret-test-297c87ac-6d23-49ac-b8ad-6670203fc668 12/14/22 16:33:03.198 - STEP: Creating a pod to test consume secrets 12/14/22 16:33:03.204 - Dec 14 16:33:03.218: INFO: Waiting up to 5m0s for pod "pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f" in namespace "secrets-4640" to be "Succeeded or Failed" - Dec 14 16:33:03.225: INFO: Pod "pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f": Phase="Pending", Reason="", readiness=false. Elapsed: 7.140027ms - Dec 14 16:33:05.231: INFO: Pod "pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013315786s - Dec 14 16:33:07.233: INFO: Pod "pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014845711s - STEP: Saw pod success 12/14/22 16:33:07.233 - Dec 14 16:33:07.233: INFO: Pod "pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f" satisfied condition "Succeeded or Failed" - Dec 14 16:33:07.238: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f container secret-volume-test: - STEP: delete the pod 12/14/22 16:33:07.249 - Dec 14 16:33:07.270: INFO: Waiting for pod pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f to disappear - Dec 14 16:33:07.279: INFO: Pod pod-secrets-b0ecfcea-7c8d-475b-8805-ee62b9ac482f no longer exists - [AfterEach] [sig-storage] Secrets + STEP: Creating a kubernetes client 02/25/23 10:37:19.241 + Feb 25 10:37:19.242: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename job 02/25/23 10:37:19.244 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:37:19.287 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:37:19.298 + [It] should adopt matching orphans and release non-matching pods [Conformance] + test/e2e/apps/job.go:335 + STEP: Creating a job 02/25/23 10:37:19.311 + STEP: Ensuring active pods == parallelism 02/25/23 10:37:19.339 + STEP: Orphaning one of the Job's Pods 02/25/23 10:37:23.348 + Feb 25 10:37:23.882: INFO: Successfully updated pod "adopt-release-5nfkf" + STEP: Checking that the Job readopts the Pod 02/25/23 10:37:23.883 + Feb 25 10:37:23.883: INFO: Waiting up to 15m0s for pod "adopt-release-5nfkf" in namespace "job-3771" to be "adopted" + Feb 25 10:37:23.904: INFO: Pod "adopt-release-5nfkf": Phase="Running", Reason="", readiness=true. Elapsed: 20.260094ms + Feb 25 10:37:25.916: INFO: Pod "adopt-release-5nfkf": Phase="Running", Reason="", readiness=true. Elapsed: 2.032376609s + Feb 25 10:37:25.916: INFO: Pod "adopt-release-5nfkf" satisfied condition "adopted" + STEP: Removing the labels from the Job's Pod 02/25/23 10:37:25.916 + Feb 25 10:37:26.447: INFO: Successfully updated pod "adopt-release-5nfkf" + STEP: Checking that the Job releases the Pod 02/25/23 10:37:26.447 + Feb 25 10:37:26.447: INFO: Waiting up to 15m0s for pod "adopt-release-5nfkf" in namespace "job-3771" to be "released" + Feb 25 10:37:26.465: INFO: Pod "adopt-release-5nfkf": Phase="Running", Reason="", readiness=true. Elapsed: 18.024647ms + Feb 25 10:37:28.477: INFO: Pod "adopt-release-5nfkf": Phase="Running", Reason="", readiness=true. Elapsed: 2.030329528s + Feb 25 10:37:28.477: INFO: Pod "adopt-release-5nfkf" satisfied condition "released" + [AfterEach] [sig-apps] Job test/e2e/framework/framework.go:187 - Dec 14 16:33:07.280: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "secrets-4640" for this suite. 12/14/22 16:33:07.285 + Feb 25 10:37:28.478: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "job-3771" for this suite. 02/25/23 10:37:28.493 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:216 -[BeforeEach] [sig-storage] EmptyDir volumes +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + works for CRD without validation schema [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:152 +[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:33:07.311 -Dec 14 16:33:07.312: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir 12/14/22 16:33:07.313 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:33:07.354 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:33:07.358 -[It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:216 -STEP: Creating a pod to test emptydir 0777 on node default medium 12/14/22 16:33:07.361 -Dec 14 16:33:07.372: INFO: Waiting up to 5m0s for pod "pod-251fb505-80a4-4149-a0c7-b94d531f77df" in namespace "emptydir-3951" to be "Succeeded or Failed" -Dec 14 16:33:07.386: INFO: Pod "pod-251fb505-80a4-4149-a0c7-b94d531f77df": Phase="Pending", Reason="", readiness=false. Elapsed: 13.923379ms -Dec 14 16:33:09.397: INFO: Pod "pod-251fb505-80a4-4149-a0c7-b94d531f77df": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025592899s -Dec 14 16:33:11.394: INFO: Pod "pod-251fb505-80a4-4149-a0c7-b94d531f77df": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021915584s -STEP: Saw pod success 12/14/22 16:33:11.394 -Dec 14 16:33:11.394: INFO: Pod "pod-251fb505-80a4-4149-a0c7-b94d531f77df" satisfied condition "Succeeded or Failed" -Dec 14 16:33:11.405: INFO: Trying to get logs from node kiedai4iakae-1 pod pod-251fb505-80a4-4149-a0c7-b94d531f77df container test-container: -STEP: delete the pod 12/14/22 16:33:11.435 -Dec 14 16:33:11.454: INFO: Waiting for pod pod-251fb505-80a4-4149-a0c7-b94d531f77df to disappear -Dec 14 16:33:11.459: INFO: Pod pod-251fb505-80a4-4149-a0c7-b94d531f77df no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +STEP: Creating a kubernetes client 02/25/23 10:37:28.546 +Feb 25 10:37:28.547: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 10:37:28.553 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:37:28.602 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:37:28.61 +[It] works for CRD without validation schema [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:152 +Feb 25 10:37:28.616: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: kubectl validation (kubectl create and apply) allows request with any unknown properties 02/25/23 10:37:37.618 +Feb 25 10:37:37.619: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-1755 --namespace=crd-publish-openapi-1755 create -f -' +Feb 25 10:37:39.635: INFO: stderr: "" +Feb 25 10:37:39.635: INFO: stdout: "e2e-test-crd-publish-openapi-5604-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" +Feb 25 10:37:39.647: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-1755 --namespace=crd-publish-openapi-1755 delete e2e-test-crd-publish-openapi-5604-crds test-cr' +Feb 25 10:37:39.992: INFO: stderr: "" +Feb 25 10:37:39.992: INFO: stdout: "e2e-test-crd-publish-openapi-5604-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" +Feb 25 10:37:39.994: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-1755 --namespace=crd-publish-openapi-1755 apply -f -' +Feb 25 10:37:41.607: INFO: stderr: "" +Feb 25 10:37:41.607: INFO: stdout: "e2e-test-crd-publish-openapi-5604-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" +Feb 25 10:37:41.608: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-1755 --namespace=crd-publish-openapi-1755 delete e2e-test-crd-publish-openapi-5604-crds test-cr' +Feb 25 10:37:41.871: INFO: stderr: "" +Feb 25 10:37:41.871: INFO: stdout: "e2e-test-crd-publish-openapi-5604-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" +STEP: kubectl explain works to explain CR without validation schema 02/25/23 10:37:41.872 +Feb 25 10:37:41.875: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-1755 explain e2e-test-crd-publish-openapi-5604-crds' +Feb 25 10:37:42.688: INFO: stderr: "" +Feb 25 10:37:42.688: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-5604-crd\nVERSION: crd-publish-openapi-test-empty.example.com/v1\n\nDESCRIPTION:\n \n" +[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:33:11.459: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-3951" for this suite. 12/14/22 16:33:11.465 -{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","completed":170,"skipped":3076,"failed":0} +Feb 25 10:37:50.727: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "crd-publish-openapi-1755" for this suite. 02/25/23 10:37:50.749 +{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD without validation schema [Conformance]","completed":168,"skipped":3184,"failed":0} ------------------------------ -• [4.165 seconds] -[sig-storage] EmptyDir volumes -test/e2e/common/storage/framework.go:23 - should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:216 +• [SLOW TEST] [22.219 seconds] +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + works for CRD without validation schema [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:152 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir volumes + [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:33:07.311 - Dec 14 16:33:07.312: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir 12/14/22 16:33:07.313 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:33:07.354 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:33:07.358 - [It] should support (non-root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:216 - STEP: Creating a pod to test emptydir 0777 on node default medium 12/14/22 16:33:07.361 - Dec 14 16:33:07.372: INFO: Waiting up to 5m0s for pod "pod-251fb505-80a4-4149-a0c7-b94d531f77df" in namespace "emptydir-3951" to be "Succeeded or Failed" - Dec 14 16:33:07.386: INFO: Pod "pod-251fb505-80a4-4149-a0c7-b94d531f77df": Phase="Pending", Reason="", readiness=false. Elapsed: 13.923379ms - Dec 14 16:33:09.397: INFO: Pod "pod-251fb505-80a4-4149-a0c7-b94d531f77df": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025592899s - Dec 14 16:33:11.394: INFO: Pod "pod-251fb505-80a4-4149-a0c7-b94d531f77df": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021915584s - STEP: Saw pod success 12/14/22 16:33:11.394 - Dec 14 16:33:11.394: INFO: Pod "pod-251fb505-80a4-4149-a0c7-b94d531f77df" satisfied condition "Succeeded or Failed" - Dec 14 16:33:11.405: INFO: Trying to get logs from node kiedai4iakae-1 pod pod-251fb505-80a4-4149-a0c7-b94d531f77df container test-container: - STEP: delete the pod 12/14/22 16:33:11.435 - Dec 14 16:33:11.454: INFO: Waiting for pod pod-251fb505-80a4-4149-a0c7-b94d531f77df to disappear - Dec 14 16:33:11.459: INFO: Pod pod-251fb505-80a4-4149-a0c7-b94d531f77df no longer exists - [AfterEach] [sig-storage] EmptyDir volumes + STEP: Creating a kubernetes client 02/25/23 10:37:28.546 + Feb 25 10:37:28.547: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 10:37:28.553 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:37:28.602 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:37:28.61 + [It] works for CRD without validation schema [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:152 + Feb 25 10:37:28.616: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: kubectl validation (kubectl create and apply) allows request with any unknown properties 02/25/23 10:37:37.618 + Feb 25 10:37:37.619: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-1755 --namespace=crd-publish-openapi-1755 create -f -' + Feb 25 10:37:39.635: INFO: stderr: "" + Feb 25 10:37:39.635: INFO: stdout: "e2e-test-crd-publish-openapi-5604-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" + Feb 25 10:37:39.647: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-1755 --namespace=crd-publish-openapi-1755 delete e2e-test-crd-publish-openapi-5604-crds test-cr' + Feb 25 10:37:39.992: INFO: stderr: "" + Feb 25 10:37:39.992: INFO: stdout: "e2e-test-crd-publish-openapi-5604-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" + Feb 25 10:37:39.994: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-1755 --namespace=crd-publish-openapi-1755 apply -f -' + Feb 25 10:37:41.607: INFO: stderr: "" + Feb 25 10:37:41.607: INFO: stdout: "e2e-test-crd-publish-openapi-5604-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" + Feb 25 10:37:41.608: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-1755 --namespace=crd-publish-openapi-1755 delete e2e-test-crd-publish-openapi-5604-crds test-cr' + Feb 25 10:37:41.871: INFO: stderr: "" + Feb 25 10:37:41.871: INFO: stdout: "e2e-test-crd-publish-openapi-5604-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" + STEP: kubectl explain works to explain CR without validation schema 02/25/23 10:37:41.872 + Feb 25 10:37:41.875: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-1755 explain e2e-test-crd-publish-openapi-5604-crds' + Feb 25 10:37:42.688: INFO: stderr: "" + Feb 25 10:37:42.688: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-5604-crd\nVERSION: crd-publish-openapi-test-empty.example.com/v1\n\nDESCRIPTION:\n \n" + [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:33:11.459: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-3951" for this suite. 12/14/22 16:33:11.465 + Feb 25 10:37:50.727: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "crd-publish-openapi-1755" for this suite. 02/25/23 10:37:50.749 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - should include webhook resources in discovery documents [Conformance] - test/e2e/apimachinery/webhook.go:116 + should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] + test/e2e/apimachinery/webhook.go:276 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:33:11.482 -Dec 14 16:33:11.482: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 16:33:11.486 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:33:11.509 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:33:11.511 +STEP: Creating a kubernetes client 02/25/23 10:37:50.771 +Feb 25 10:37:50.772: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 10:37:50.774 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:37:50.807 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:37:50.815 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 16:33:11.534 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:33:12.05 -STEP: Deploying the webhook pod 12/14/22 16:33:12.059 -STEP: Wait for the deployment to be ready 12/14/22 16:33:12.074 -Dec 14 16:33:12.099: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set -STEP: Deploying the webhook service 12/14/22 16:33:14.119 -STEP: Verifying the service has paired with the endpoint 12/14/22 16:33:14.151 -Dec 14 16:33:15.151: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] should include webhook resources in discovery documents [Conformance] - test/e2e/apimachinery/webhook.go:116 -STEP: fetching the /apis discovery document 12/14/22 16:33:15.166 -STEP: finding the admissionregistration.k8s.io API group in the /apis discovery document 12/14/22 16:33:15.169 -STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis discovery document 12/14/22 16:33:15.169 -STEP: fetching the /apis/admissionregistration.k8s.io discovery document 12/14/22 16:33:15.169 -STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis/admissionregistration.k8s.io discovery document 12/14/22 16:33:15.174 -STEP: fetching the /apis/admissionregistration.k8s.io/v1 discovery document 12/14/22 16:33:15.174 -STEP: finding mutatingwebhookconfigurations and validatingwebhookconfigurations resources in the /apis/admissionregistration.k8s.io/v1 discovery document 12/14/22 16:33:15.175 +STEP: Setting up server cert 02/25/23 10:37:50.852 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:37:52.477 +STEP: Deploying the webhook pod 02/25/23 10:37:52.499 +STEP: Wait for the deployment to be ready 02/25/23 10:37:52.528 +Feb 25 10:37:52.550: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set +STEP: Deploying the webhook service 02/25/23 10:37:54.594 +STEP: Verifying the service has paired with the endpoint 02/25/23 10:37:54.629 +Feb 25 10:37:55.630: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] + test/e2e/apimachinery/webhook.go:276 +STEP: Registering a validating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API 02/25/23 10:37:55.64 +STEP: Registering a mutating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API 02/25/23 10:37:55.692 +STEP: Creating a dummy validating-webhook-configuration object 02/25/23 10:37:55.741 +STEP: Deleting the validating-webhook-configuration, which should be possible to remove 02/25/23 10:37:55.762 +STEP: Creating a dummy mutating-webhook-configuration object 02/25/23 10:37:55.783 +STEP: Deleting the mutating-webhook-configuration, which should be possible to remove 02/25/23 10:37:55.806 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:33:15.176: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-954" for this suite. 12/14/22 16:33:15.182 -STEP: Destroying namespace "webhook-954-markers" for this suite. 12/14/22 16:33:15.197 +Feb 25 10:37:55.858: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-6599" for this suite. 02/25/23 10:37:55.874 +STEP: Destroying namespace "webhook-6599-markers" for this suite. 02/25/23 10:37:55.9 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should include webhook resources in discovery documents [Conformance]","completed":171,"skipped":3103,"failed":0} +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]","completed":169,"skipped":3195,"failed":0} ------------------------------ -• [3.777 seconds] +• [SLOW TEST] [5.324 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/framework.go:23 - should include webhook resources in discovery documents [Conformance] - test/e2e/apimachinery/webhook.go:116 + should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] + test/e2e/apimachinery/webhook.go:276 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:33:11.482 - Dec 14 16:33:11.482: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 16:33:11.486 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:33:11.509 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:33:11.511 + STEP: Creating a kubernetes client 02/25/23 10:37:50.771 + Feb 25 10:37:50.772: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 10:37:50.774 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:37:50.807 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:37:50.815 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 16:33:11.534 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:33:12.05 - STEP: Deploying the webhook pod 12/14/22 16:33:12.059 - STEP: Wait for the deployment to be ready 12/14/22 16:33:12.074 - Dec 14 16:33:12.099: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set - STEP: Deploying the webhook service 12/14/22 16:33:14.119 - STEP: Verifying the service has paired with the endpoint 12/14/22 16:33:14.151 - Dec 14 16:33:15.151: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] should include webhook resources in discovery documents [Conformance] - test/e2e/apimachinery/webhook.go:116 - STEP: fetching the /apis discovery document 12/14/22 16:33:15.166 - STEP: finding the admissionregistration.k8s.io API group in the /apis discovery document 12/14/22 16:33:15.169 - STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis discovery document 12/14/22 16:33:15.169 - STEP: fetching the /apis/admissionregistration.k8s.io discovery document 12/14/22 16:33:15.169 - STEP: finding the admissionregistration.k8s.io/v1 API group/version in the /apis/admissionregistration.k8s.io discovery document 12/14/22 16:33:15.174 - STEP: fetching the /apis/admissionregistration.k8s.io/v1 discovery document 12/14/22 16:33:15.174 - STEP: finding mutatingwebhookconfigurations and validatingwebhookconfigurations resources in the /apis/admissionregistration.k8s.io/v1 discovery document 12/14/22 16:33:15.175 + STEP: Setting up server cert 02/25/23 10:37:50.852 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:37:52.477 + STEP: Deploying the webhook pod 02/25/23 10:37:52.499 + STEP: Wait for the deployment to be ready 02/25/23 10:37:52.528 + Feb 25 10:37:52.550: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set + STEP: Deploying the webhook service 02/25/23 10:37:54.594 + STEP: Verifying the service has paired with the endpoint 02/25/23 10:37:54.629 + Feb 25 10:37:55.630: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] + test/e2e/apimachinery/webhook.go:276 + STEP: Registering a validating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API 02/25/23 10:37:55.64 + STEP: Registering a mutating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API 02/25/23 10:37:55.692 + STEP: Creating a dummy validating-webhook-configuration object 02/25/23 10:37:55.741 + STEP: Deleting the validating-webhook-configuration, which should be possible to remove 02/25/23 10:37:55.762 + STEP: Creating a dummy mutating-webhook-configuration object 02/25/23 10:37:55.783 + STEP: Deleting the mutating-webhook-configuration, which should be possible to remove 02/25/23 10:37:55.806 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:33:15.176: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-954" for this suite. 12/14/22 16:33:15.182 - STEP: Destroying namespace "webhook-954-markers" for this suite. 12/14/22 16:33:15.197 + Feb 25 10:37:55.858: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-6599" for this suite. 02/25/23 10:37:55.874 + STEP: Destroying namespace "webhook-6599-markers" for this suite. 02/25/23 10:37:55.9 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:104 << End Captured GinkgoWriter Output ------------------------------ -SSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Pods - should delete a collection of pods [Conformance] - test/e2e/common/node/pods.go:844 -[BeforeEach] [sig-node] Pods +[sig-storage] Downward API volume + should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:248 +[BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:33:15.261 -Dec 14 16:33:15.263: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename pods 12/14/22 16:33:15.267 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:33:15.309 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:33:15.313 -[BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 -[It] should delete a collection of pods [Conformance] - test/e2e/common/node/pods.go:844 -STEP: Create set of pods 12/14/22 16:33:15.319 -Dec 14 16:33:15.334: INFO: created test-pod-1 -Dec 14 16:33:15.342: INFO: created test-pod-2 -Dec 14 16:33:15.371: INFO: created test-pod-3 -STEP: waiting for all 3 pods to be running 12/14/22 16:33:15.371 -Dec 14 16:33:15.372: INFO: Waiting up to 5m0s for all pods (need at least 3) in namespace 'pods-3837' to be running and ready -Dec 14 16:33:15.410: INFO: The status of Pod test-pod-1 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed -Dec 14 16:33:15.410: INFO: The status of Pod test-pod-2 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed -Dec 14 16:33:15.410: INFO: The status of Pod test-pod-3 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed -Dec 14 16:33:15.410: INFO: 0 / 3 pods in namespace 'pods-3837' are running and ready (0 seconds elapsed) -Dec 14 16:33:15.410: INFO: expected 0 pod replicas in namespace 'pods-3837', 0 are Running and Ready. -Dec 14 16:33:15.410: INFO: POD NODE PHASE GRACE CONDITIONS -Dec 14 16:33:15.410: INFO: test-pod-1 kiedai4iakae-3 Pending [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC }] -Dec 14 16:33:15.410: INFO: test-pod-2 kiedai4iakae-3 Pending [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC }] -Dec 14 16:33:15.410: INFO: test-pod-3 kiedai4iakae-3 Pending [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC }] -Dec 14 16:33:15.410: INFO: -Dec 14 16:33:17.456: INFO: The status of Pod test-pod-1 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed -Dec 14 16:33:17.456: INFO: The status of Pod test-pod-2 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed -Dec 14 16:33:17.457: INFO: The status of Pod test-pod-3 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed -Dec 14 16:33:17.457: INFO: 0 / 3 pods in namespace 'pods-3837' are running and ready (2 seconds elapsed) -Dec 14 16:33:17.457: INFO: expected 0 pod replicas in namespace 'pods-3837', 0 are Running and Ready. -Dec 14 16:33:17.457: INFO: POD NODE PHASE GRACE CONDITIONS -Dec 14 16:33:17.457: INFO: test-pod-1 kiedai4iakae-3 Pending [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC }] -Dec 14 16:33:17.457: INFO: test-pod-2 kiedai4iakae-3 Pending [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC }] -Dec 14 16:33:17.458: INFO: test-pod-3 kiedai4iakae-3 Pending [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC }] -Dec 14 16:33:17.458: INFO: -Dec 14 16:33:19.428: INFO: 3 / 3 pods in namespace 'pods-3837' are running and ready (4 seconds elapsed) -Dec 14 16:33:19.428: INFO: expected 0 pod replicas in namespace 'pods-3837', 0 are Running and Ready. -STEP: waiting for all pods to be deleted 12/14/22 16:33:19.506 -Dec 14 16:33:19.512: INFO: Pod quantity 3 is different from expected quantity 0 -Dec 14 16:33:20.519: INFO: Pod quantity 3 is different from expected quantity 0 -[AfterEach] [sig-node] Pods +STEP: Creating a kubernetes client 02/25/23 10:37:56.132 +Feb 25 10:37:56.132: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 10:37:56.14 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:37:56.212 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:37:56.225 +[BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 +[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:248 +STEP: Creating a pod to test downward API volume plugin 02/25/23 10:37:56.239 +Feb 25 10:37:56.266: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21" in namespace "downward-api-961" to be "Succeeded or Failed" +Feb 25 10:37:56.276: INFO: Pod "downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21": Phase="Pending", Reason="", readiness=false. Elapsed: 9.926588ms +Feb 25 10:37:58.285: INFO: Pod "downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01933871s +Feb 25 10:38:00.290: INFO: Pod "downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21": Phase="Pending", Reason="", readiness=false. Elapsed: 4.024220473s +Feb 25 10:38:02.286: INFO: Pod "downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.020338558s +STEP: Saw pod success 02/25/23 10:38:02.286 +Feb 25 10:38:02.287: INFO: Pod "downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21" satisfied condition "Succeeded or Failed" +Feb 25 10:38:02.295: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21 container client-container: +STEP: delete the pod 02/25/23 10:38:02.326 +Feb 25 10:38:02.357: INFO: Waiting for pod downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21 to disappear +Feb 25 10:38:02.368: INFO: Pod downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21 no longer exists +[AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 -Dec 14 16:33:21.522: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-3837" for this suite. 12/14/22 16:33:21.531 -{"msg":"PASSED [sig-node] Pods should delete a collection of pods [Conformance]","completed":172,"skipped":3109,"failed":0} +Feb 25 10:38:02.369: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-961" for this suite. 02/25/23 10:38:02.387 +{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","completed":170,"skipped":3240,"failed":0} ------------------------------ • [SLOW TEST] [6.280 seconds] -[sig-node] Pods -test/e2e/common/node/framework.go:23 - should delete a collection of pods [Conformance] - test/e2e/common/node/pods.go:844 +[sig-storage] Downward API volume +test/e2e/common/storage/framework.go:23 + should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:248 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Pods + [BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:33:15.261 - Dec 14 16:33:15.263: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename pods 12/14/22 16:33:15.267 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:33:15.309 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:33:15.313 - [BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 - [It] should delete a collection of pods [Conformance] - test/e2e/common/node/pods.go:844 - STEP: Create set of pods 12/14/22 16:33:15.319 - Dec 14 16:33:15.334: INFO: created test-pod-1 - Dec 14 16:33:15.342: INFO: created test-pod-2 - Dec 14 16:33:15.371: INFO: created test-pod-3 - STEP: waiting for all 3 pods to be running 12/14/22 16:33:15.371 - Dec 14 16:33:15.372: INFO: Waiting up to 5m0s for all pods (need at least 3) in namespace 'pods-3837' to be running and ready - Dec 14 16:33:15.410: INFO: The status of Pod test-pod-1 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed - Dec 14 16:33:15.410: INFO: The status of Pod test-pod-2 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed - Dec 14 16:33:15.410: INFO: The status of Pod test-pod-3 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed - Dec 14 16:33:15.410: INFO: 0 / 3 pods in namespace 'pods-3837' are running and ready (0 seconds elapsed) - Dec 14 16:33:15.410: INFO: expected 0 pod replicas in namespace 'pods-3837', 0 are Running and Ready. - Dec 14 16:33:15.410: INFO: POD NODE PHASE GRACE CONDITIONS - Dec 14 16:33:15.410: INFO: test-pod-1 kiedai4iakae-3 Pending [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC }] - Dec 14 16:33:15.410: INFO: test-pod-2 kiedai4iakae-3 Pending [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC }] - Dec 14 16:33:15.410: INFO: test-pod-3 kiedai4iakae-3 Pending [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC }] - Dec 14 16:33:15.410: INFO: - Dec 14 16:33:17.456: INFO: The status of Pod test-pod-1 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed - Dec 14 16:33:17.456: INFO: The status of Pod test-pod-2 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed - Dec 14 16:33:17.457: INFO: The status of Pod test-pod-3 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed - Dec 14 16:33:17.457: INFO: 0 / 3 pods in namespace 'pods-3837' are running and ready (2 seconds elapsed) - Dec 14 16:33:17.457: INFO: expected 0 pod replicas in namespace 'pods-3837', 0 are Running and Ready. - Dec 14 16:33:17.457: INFO: POD NODE PHASE GRACE CONDITIONS - Dec 14 16:33:17.457: INFO: test-pod-1 kiedai4iakae-3 Pending [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC }] - Dec 14 16:33:17.457: INFO: test-pod-2 kiedai4iakae-3 Pending [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC }] - Dec 14 16:33:17.458: INFO: test-pod-3 kiedai4iakae-3 Pending [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC ContainersNotReady containers with unready status: [token-test]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:33:15 +0000 UTC }] - Dec 14 16:33:17.458: INFO: - Dec 14 16:33:19.428: INFO: 3 / 3 pods in namespace 'pods-3837' are running and ready (4 seconds elapsed) - Dec 14 16:33:19.428: INFO: expected 0 pod replicas in namespace 'pods-3837', 0 are Running and Ready. - STEP: waiting for all pods to be deleted 12/14/22 16:33:19.506 - Dec 14 16:33:19.512: INFO: Pod quantity 3 is different from expected quantity 0 - Dec 14 16:33:20.519: INFO: Pod quantity 3 is different from expected quantity 0 - [AfterEach] [sig-node] Pods + STEP: Creating a kubernetes client 02/25/23 10:37:56.132 + Feb 25 10:37:56.132: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 10:37:56.14 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:37:56.212 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:37:56.225 + [BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 + [It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:248 + STEP: Creating a pod to test downward API volume plugin 02/25/23 10:37:56.239 + Feb 25 10:37:56.266: INFO: Waiting up to 5m0s for pod "downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21" in namespace "downward-api-961" to be "Succeeded or Failed" + Feb 25 10:37:56.276: INFO: Pod "downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21": Phase="Pending", Reason="", readiness=false. Elapsed: 9.926588ms + Feb 25 10:37:58.285: INFO: Pod "downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01933871s + Feb 25 10:38:00.290: INFO: Pod "downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21": Phase="Pending", Reason="", readiness=false. Elapsed: 4.024220473s + Feb 25 10:38:02.286: INFO: Pod "downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.020338558s + STEP: Saw pod success 02/25/23 10:38:02.286 + Feb 25 10:38:02.287: INFO: Pod "downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21" satisfied condition "Succeeded or Failed" + Feb 25 10:38:02.295: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21 container client-container: + STEP: delete the pod 02/25/23 10:38:02.326 + Feb 25 10:38:02.357: INFO: Waiting for pod downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21 to disappear + Feb 25 10:38:02.368: INFO: Pod downwardapi-volume-f0452176-a0f2-40cd-b4e6-0ba8e0918a21 no longer exists + [AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 - Dec 14 16:33:21.522: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "pods-3837" for this suite. 12/14/22 16:33:21.531 + Feb 25 10:38:02.369: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-961" for this suite. 02/25/23 10:38:02.387 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSS +S ------------------------------ -[sig-network] DNS - should provide DNS for services [Conformance] - test/e2e/network/dns.go:137 -[BeforeEach] [sig-network] DNS +[sig-api-machinery] Garbage collector + should orphan pods created by rc if delete options say so [Conformance] + test/e2e/apimachinery/garbage_collector.go:370 +[BeforeEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:33:21.549 -Dec 14 16:33:21.549: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename dns 12/14/22 16:33:21.551 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:33:21.58 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:33:21.583 -[It] should provide DNS for services [Conformance] - test/e2e/network/dns.go:137 -STEP: Creating a test headless service 12/14/22 16:33:21.587 -STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-4802.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-4802.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-4802.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-4802.svc.cluster.local;check="$$(dig +notcp +noall +answer +search 91.48.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.48.91_udp@PTR;check="$$(dig +tcp +noall +answer +search 91.48.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.48.91_tcp@PTR;sleep 1; done - 12/14/22 16:33:21.612 -STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-4802.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-4802.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-4802.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-4802.svc.cluster.local;check="$$(dig +notcp +noall +answer +search 91.48.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.48.91_udp@PTR;check="$$(dig +tcp +noall +answer +search 91.48.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.48.91_tcp@PTR;sleep 1; done - 12/14/22 16:33:21.613 -STEP: creating a pod to probe DNS 12/14/22 16:33:21.613 -STEP: submitting the pod to kubernetes 12/14/22 16:33:21.613 -Dec 14 16:33:21.634: INFO: Waiting up to 15m0s for pod "dns-test-ee547afa-b447-4422-a826-cc33a77b36fc" in namespace "dns-4802" to be "running" -Dec 14 16:33:21.641: INFO: Pod "dns-test-ee547afa-b447-4422-a826-cc33a77b36fc": Phase="Pending", Reason="", readiness=false. Elapsed: 6.65757ms -Dec 14 16:33:23.652: INFO: Pod "dns-test-ee547afa-b447-4422-a826-cc33a77b36fc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017879072s -Dec 14 16:33:25.651: INFO: Pod "dns-test-ee547afa-b447-4422-a826-cc33a77b36fc": Phase="Running", Reason="", readiness=true. Elapsed: 4.016561579s -Dec 14 16:33:25.651: INFO: Pod "dns-test-ee547afa-b447-4422-a826-cc33a77b36fc" satisfied condition "running" -STEP: retrieving the pod 12/14/22 16:33:25.651 -STEP: looking for the results for each expected name from probers 12/14/22 16:33:25.658 -Dec 14 16:33:25.671: INFO: Unable to read wheezy_tcp@dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:25.676: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:25.682: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:25.720: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:25.726: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:25.748: INFO: Lookups using dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc failed for: [wheezy_tcp@dns-test-service.dns-4802.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local] - -Dec 14 16:33:30.779: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:30.783: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:30.825: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:30.831: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:30.855: INFO: Lookups using dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc failed for: [wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local] - -Dec 14 16:33:35.787: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:35.794: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:35.852: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:35.857: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:35.904: INFO: Lookups using dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc failed for: [wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local] - -Dec 14 16:33:40.776: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:40.781: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:40.817: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:40.822: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:40.844: INFO: Lookups using dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc failed for: [wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local] - -Dec 14 16:33:45.766: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:45.771: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:45.826: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:45.832: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:45.874: INFO: Lookups using dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc failed for: [wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local] - -Dec 14 16:33:50.770: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:50.776: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:50.810: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:50.814: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) -Dec 14 16:33:50.833: INFO: Lookups using dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc failed for: [wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local] - -Dec 14 16:33:55.920: INFO: DNS probes using dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc succeeded - -STEP: deleting the pod 12/14/22 16:33:55.92 -STEP: deleting the test service 12/14/22 16:33:55.937 -STEP: deleting the test headless service 12/14/22 16:33:55.964 -[AfterEach] [sig-network] DNS - test/e2e/framework/framework.go:187 -Dec 14 16:33:55.981: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "dns-4802" for this suite. 12/14/22 16:33:55.992 -{"msg":"PASSED [sig-network] DNS should provide DNS for services [Conformance]","completed":173,"skipped":3129,"failed":0} ------------------------------- -• [SLOW TEST] [34.457 seconds] -[sig-network] DNS -test/e2e/network/common/framework.go:23 - should provide DNS for services [Conformance] - test/e2e/network/dns.go:137 +STEP: Creating a kubernetes client 02/25/23 10:38:02.418 +Feb 25 10:38:02.418: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename gc 02/25/23 10:38:02.424 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:38:02.469 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:38:02.476 +[It] should orphan pods created by rc if delete options say so [Conformance] + test/e2e/apimachinery/garbage_collector.go:370 +STEP: create the rc 02/25/23 10:38:02.492 +STEP: delete the rc 02/25/23 10:38:07.519 +STEP: wait for the rc to be deleted 02/25/23 10:38:07.675 +STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods 02/25/23 10:38:12.694 +STEP: Gathering metrics 02/25/23 10:38:43.149 +Feb 25 10:38:43.710: INFO: Waiting up to 5m0s for pod "kube-controller-manager-diek3dua9gay-2" in namespace "kube-system" to be "running and ready" +Feb 25 10:38:43.717: INFO: Pod "kube-controller-manager-diek3dua9gay-2": Phase="Running", Reason="", readiness=true. Elapsed: 6.760534ms +Feb 25 10:38:43.717: INFO: The phase of Pod kube-controller-manager-diek3dua9gay-2 is Running (Ready = true) +Feb 25 10:38:43.717: INFO: Pod "kube-controller-manager-diek3dua9gay-2" satisfied condition "running and ready" +Feb 25 10:38:43.886: INFO: For apiserver_request_total: +For apiserver_request_latency_seconds: +For apiserver_init_events_total: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] DNS - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:33:21.549 - Dec 14 16:33:21.549: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename dns 12/14/22 16:33:21.551 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:33:21.58 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:33:21.583 - [It] should provide DNS for services [Conformance] - test/e2e/network/dns.go:137 - STEP: Creating a test headless service 12/14/22 16:33:21.587 - STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-4802.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-4802.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-4802.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-4802.svc.cluster.local;check="$$(dig +notcp +noall +answer +search 91.48.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.48.91_udp@PTR;check="$$(dig +tcp +noall +answer +search 91.48.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.48.91_tcp@PTR;sleep 1; done - 12/14/22 16:33:21.612 - STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-4802.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-4802.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-4802.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-4802.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-4802.svc.cluster.local;check="$$(dig +notcp +noall +answer +search 91.48.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.48.91_udp@PTR;check="$$(dig +tcp +noall +answer +search 91.48.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.48.91_tcp@PTR;sleep 1; done - 12/14/22 16:33:21.613 - STEP: creating a pod to probe DNS 12/14/22 16:33:21.613 - STEP: submitting the pod to kubernetes 12/14/22 16:33:21.613 - Dec 14 16:33:21.634: INFO: Waiting up to 15m0s for pod "dns-test-ee547afa-b447-4422-a826-cc33a77b36fc" in namespace "dns-4802" to be "running" - Dec 14 16:33:21.641: INFO: Pod "dns-test-ee547afa-b447-4422-a826-cc33a77b36fc": Phase="Pending", Reason="", readiness=false. Elapsed: 6.65757ms - Dec 14 16:33:23.652: INFO: Pod "dns-test-ee547afa-b447-4422-a826-cc33a77b36fc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017879072s - Dec 14 16:33:25.651: INFO: Pod "dns-test-ee547afa-b447-4422-a826-cc33a77b36fc": Phase="Running", Reason="", readiness=true. Elapsed: 4.016561579s - Dec 14 16:33:25.651: INFO: Pod "dns-test-ee547afa-b447-4422-a826-cc33a77b36fc" satisfied condition "running" - STEP: retrieving the pod 12/14/22 16:33:25.651 - STEP: looking for the results for each expected name from probers 12/14/22 16:33:25.658 - Dec 14 16:33:25.671: INFO: Unable to read wheezy_tcp@dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:25.676: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:25.682: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:25.720: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:25.726: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:25.748: INFO: Lookups using dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc failed for: [wheezy_tcp@dns-test-service.dns-4802.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local] - - Dec 14 16:33:30.779: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:30.783: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:30.825: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:30.831: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:30.855: INFO: Lookups using dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc failed for: [wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local] - - Dec 14 16:33:35.787: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:35.794: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:35.852: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:35.857: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:35.904: INFO: Lookups using dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc failed for: [wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local] - - Dec 14 16:33:40.776: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:40.781: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:40.817: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:40.822: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:40.844: INFO: Lookups using dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc failed for: [wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local] - - Dec 14 16:33:45.766: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:45.771: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:45.826: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:45.832: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:45.874: INFO: Lookups using dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc failed for: [wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local] - - Dec 14 16:33:50.770: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:50.776: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:50.810: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:50.814: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local from pod dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc: the server could not find the requested resource (get pods dns-test-ee547afa-b447-4422-a826-cc33a77b36fc) - Dec 14 16:33:50.833: INFO: Lookups using dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc failed for: [wheezy_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-4802.svc.cluster.local] - - Dec 14 16:33:55.920: INFO: DNS probes using dns-4802/dns-test-ee547afa-b447-4422-a826-cc33a77b36fc succeeded - - STEP: deleting the pod 12/14/22 16:33:55.92 - STEP: deleting the test service 12/14/22 16:33:55.937 - STEP: deleting the test headless service 12/14/22 16:33:55.964 - [AfterEach] [sig-network] DNS - test/e2e/framework/framework.go:187 - Dec 14 16:33:55.981: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "dns-4802" for this suite. 12/14/22 16:33:55.992 - << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSSSS ------------------------------- -[sig-storage] Secrets - should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:67 -[BeforeEach] [sig-storage] Secrets - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:33:56.008 -Dec 14 16:33:56.008: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename secrets 12/14/22 16:33:56.031 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:33:56.107 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:33:56.116 -[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:67 -STEP: Creating secret with name secret-test-f694f753-3e75-4e66-8fb5-e7e9243f5200 12/14/22 16:33:56.12 -STEP: Creating a pod to test consume secrets 12/14/22 16:33:56.135 -Dec 14 16:33:56.149: INFO: Waiting up to 5m0s for pod "pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3" in namespace "secrets-7046" to be "Succeeded or Failed" -Dec 14 16:33:56.163: INFO: Pod "pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3": Phase="Pending", Reason="", readiness=false. Elapsed: 13.642254ms -Dec 14 16:33:58.173: INFO: Pod "pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023490789s -Dec 14 16:34:00.170: INFO: Pod "pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021094448s -STEP: Saw pod success 12/14/22 16:34:00.17 -Dec 14 16:34:00.171: INFO: Pod "pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3" satisfied condition "Succeeded or Failed" -Dec 14 16:34:00.175: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3 container secret-volume-test: -STEP: delete the pod 12/14/22 16:34:00.185 -Dec 14 16:34:00.212: INFO: Waiting for pod pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3 to disappear -Dec 14 16:34:00.219: INFO: Pod pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3 no longer exists -[AfterEach] [sig-storage] Secrets +Feb 25 10:38:43.903: INFO: Deleting pod "simpletest.rc-2jbch" in namespace "gc-2058" +Feb 25 10:38:44.055: INFO: Deleting pod "simpletest.rc-2jjdn" in namespace "gc-2058" +Feb 25 10:38:44.206: INFO: Deleting pod "simpletest.rc-2vw46" in namespace "gc-2058" +Feb 25 10:38:44.347: INFO: Deleting pod "simpletest.rc-2zcch" in namespace "gc-2058" +Feb 25 10:38:44.422: INFO: Deleting pod "simpletest.rc-4hcl9" in namespace "gc-2058" +Feb 25 10:38:44.519: INFO: Deleting pod "simpletest.rc-4lvpz" in namespace "gc-2058" +Feb 25 10:38:44.783: INFO: Deleting pod "simpletest.rc-5467n" in namespace "gc-2058" +Feb 25 10:38:45.774: INFO: Deleting pod "simpletest.rc-5s9xb" in namespace "gc-2058" +Feb 25 10:38:46.400: INFO: Deleting pod "simpletest.rc-66vp6" in namespace "gc-2058" +Feb 25 10:38:46.502: INFO: Deleting pod "simpletest.rc-6j6df" in namespace "gc-2058" +Feb 25 10:38:46.594: INFO: Deleting pod "simpletest.rc-6kd5m" in namespace "gc-2058" +Feb 25 10:38:46.661: INFO: Deleting pod "simpletest.rc-6s54n" in namespace "gc-2058" +Feb 25 10:38:46.776: INFO: Deleting pod "simpletest.rc-6w59b" in namespace "gc-2058" +Feb 25 10:38:47.084: INFO: Deleting pod "simpletest.rc-7ffrx" in namespace "gc-2058" +Feb 25 10:38:47.338: INFO: Deleting pod "simpletest.rc-7h6vq" in namespace "gc-2058" +Feb 25 10:38:47.393: INFO: Deleting pod "simpletest.rc-7pr6f" in namespace "gc-2058" +Feb 25 10:38:47.466: INFO: Deleting pod "simpletest.rc-7pwx5" in namespace "gc-2058" +Feb 25 10:38:47.625: INFO: Deleting pod "simpletest.rc-7rrw4" in namespace "gc-2058" +Feb 25 10:38:47.786: INFO: Deleting pod "simpletest.rc-87pr4" in namespace "gc-2058" +Feb 25 10:38:47.901: INFO: Deleting pod "simpletest.rc-8pvw4" in namespace "gc-2058" +Feb 25 10:38:47.991: INFO: Deleting pod "simpletest.rc-8rs56" in namespace "gc-2058" +Feb 25 10:38:48.099: INFO: Deleting pod "simpletest.rc-962sn" in namespace "gc-2058" +Feb 25 10:38:48.177: INFO: Deleting pod "simpletest.rc-98982" in namespace "gc-2058" +Feb 25 10:38:48.252: INFO: Deleting pod "simpletest.rc-99xpb" in namespace "gc-2058" +Feb 25 10:38:48.426: INFO: Deleting pod "simpletest.rc-9pc9r" in namespace "gc-2058" +Feb 25 10:38:48.631: INFO: Deleting pod "simpletest.rc-b5qsf" in namespace "gc-2058" +Feb 25 10:38:48.732: INFO: Deleting pod "simpletest.rc-b687j" in namespace "gc-2058" +Feb 25 10:38:48.870: INFO: Deleting pod "simpletest.rc-bcxts" in namespace "gc-2058" +Feb 25 10:38:49.475: INFO: Deleting pod "simpletest.rc-bhnwx" in namespace "gc-2058" +Feb 25 10:38:49.866: INFO: Deleting pod "simpletest.rc-bt82t" in namespace "gc-2058" +Feb 25 10:38:50.001: INFO: Deleting pod "simpletest.rc-cbzzk" in namespace "gc-2058" +Feb 25 10:38:50.144: INFO: Deleting pod "simpletest.rc-cnkzc" in namespace "gc-2058" +Feb 25 10:38:50.310: INFO: Deleting pod "simpletest.rc-cxmjx" in namespace "gc-2058" +Feb 25 10:38:50.467: INFO: Deleting pod "simpletest.rc-d6gcj" in namespace "gc-2058" +Feb 25 10:38:50.579: INFO: Deleting pod "simpletest.rc-d8pft" in namespace "gc-2058" +Feb 25 10:38:50.671: INFO: Deleting pod "simpletest.rc-dnrb9" in namespace "gc-2058" +Feb 25 10:38:50.836: INFO: Deleting pod "simpletest.rc-drjpz" in namespace "gc-2058" +Feb 25 10:38:50.977: INFO: Deleting pod "simpletest.rc-f57nc" in namespace "gc-2058" +Feb 25 10:38:51.100: INFO: Deleting pod "simpletest.rc-f66j9" in namespace "gc-2058" +Feb 25 10:38:51.213: INFO: Deleting pod "simpletest.rc-f9cc9" in namespace "gc-2058" +Feb 25 10:38:51.277: INFO: Deleting pod "simpletest.rc-f9rwm" in namespace "gc-2058" +Feb 25 10:38:51.434: INFO: Deleting pod "simpletest.rc-ffsbr" in namespace "gc-2058" +Feb 25 10:38:51.544: INFO: Deleting pod "simpletest.rc-fqhc2" in namespace "gc-2058" +Feb 25 10:38:51.673: INFO: Deleting pod "simpletest.rc-g4td6" in namespace "gc-2058" +Feb 25 10:38:51.841: INFO: Deleting pod "simpletest.rc-gfwgm" in namespace "gc-2058" +Feb 25 10:38:51.971: INFO: Deleting pod "simpletest.rc-ghb54" in namespace "gc-2058" +Feb 25 10:38:52.084: INFO: Deleting pod "simpletest.rc-gtzxp" in namespace "gc-2058" +Feb 25 10:38:52.244: INFO: Deleting pod "simpletest.rc-hnxpc" in namespace "gc-2058" +Feb 25 10:38:52.322: INFO: Deleting pod "simpletest.rc-j2b7m" in namespace "gc-2058" +Feb 25 10:38:52.480: INFO: Deleting pod "simpletest.rc-j565f" in namespace "gc-2058" +Feb 25 10:38:52.628: INFO: Deleting pod "simpletest.rc-j5gsp" in namespace "gc-2058" +Feb 25 10:38:52.763: INFO: Deleting pod "simpletest.rc-j9r2p" in namespace "gc-2058" +Feb 25 10:38:53.177: INFO: Deleting pod "simpletest.rc-jd4jq" in namespace "gc-2058" +Feb 25 10:38:54.007: INFO: Deleting pod "simpletest.rc-jm7wv" in namespace "gc-2058" +Feb 25 10:38:54.298: INFO: Deleting pod "simpletest.rc-jvk6q" in namespace "gc-2058" +Feb 25 10:38:54.635: INFO: Deleting pod "simpletest.rc-k7gtr" in namespace "gc-2058" +Feb 25 10:38:54.864: INFO: Deleting pod "simpletest.rc-klxbd" in namespace "gc-2058" +Feb 25 10:38:55.053: INFO: Deleting pod "simpletest.rc-kvdtj" in namespace "gc-2058" +Feb 25 10:38:55.286: INFO: Deleting pod "simpletest.rc-l2pks" in namespace "gc-2058" +Feb 25 10:38:55.358: INFO: Deleting pod "simpletest.rc-lg9fl" in namespace "gc-2058" +Feb 25 10:38:55.456: INFO: Deleting pod "simpletest.rc-lrmzl" in namespace "gc-2058" +Feb 25 10:38:55.654: INFO: Deleting pod "simpletest.rc-lxtjn" in namespace "gc-2058" +Feb 25 10:38:55.981: INFO: Deleting pod "simpletest.rc-md8kp" in namespace "gc-2058" +Feb 25 10:38:56.137: INFO: Deleting pod "simpletest.rc-mgshb" in namespace "gc-2058" +Feb 25 10:38:56.275: INFO: Deleting pod "simpletest.rc-mjbj5" in namespace "gc-2058" +Feb 25 10:38:56.427: INFO: Deleting pod "simpletest.rc-n228r" in namespace "gc-2058" +Feb 25 10:38:56.587: INFO: Deleting pod "simpletest.rc-n5np6" in namespace "gc-2058" +Feb 25 10:38:56.700: INFO: Deleting pod "simpletest.rc-nhxxc" in namespace "gc-2058" +Feb 25 10:38:56.823: INFO: Deleting pod "simpletest.rc-njr4p" in namespace "gc-2058" +Feb 25 10:38:57.101: INFO: Deleting pod "simpletest.rc-nx4nd" in namespace "gc-2058" +Feb 25 10:38:58.178: INFO: Deleting pod "simpletest.rc-pntg2" in namespace "gc-2058" +Feb 25 10:38:58.381: INFO: Deleting pod "simpletest.rc-prq7h" in namespace "gc-2058" +Feb 25 10:38:58.698: INFO: Deleting pod "simpletest.rc-q99c2" in namespace "gc-2058" +Feb 25 10:38:58.807: INFO: Deleting pod "simpletest.rc-qq4cx" in namespace "gc-2058" +Feb 25 10:38:58.878: INFO: Deleting pod "simpletest.rc-qtwtp" in namespace "gc-2058" +Feb 25 10:38:59.127: INFO: Deleting pod "simpletest.rc-r55lj" in namespace "gc-2058" +Feb 25 10:38:59.286: INFO: Deleting pod "simpletest.rc-rc9kb" in namespace "gc-2058" +Feb 25 10:38:59.457: INFO: Deleting pod "simpletest.rc-rjw55" in namespace "gc-2058" +Feb 25 10:38:59.720: INFO: Deleting pod "simpletest.rc-rrq5p" in namespace "gc-2058" +Feb 25 10:38:59.870: INFO: Deleting pod "simpletest.rc-s5gth" in namespace "gc-2058" +Feb 25 10:39:00.041: INFO: Deleting pod "simpletest.rc-scqdw" in namespace "gc-2058" +Feb 25 10:39:00.108: INFO: Deleting pod "simpletest.rc-sl5fr" in namespace "gc-2058" +Feb 25 10:39:00.328: INFO: Deleting pod "simpletest.rc-srtb2" in namespace "gc-2058" +Feb 25 10:39:00.427: INFO: Deleting pod "simpletest.rc-ssv48" in namespace "gc-2058" +Feb 25 10:39:00.562: INFO: Deleting pod "simpletest.rc-swjdw" in namespace "gc-2058" +Feb 25 10:39:00.666: INFO: Deleting pod "simpletest.rc-tcth5" in namespace "gc-2058" +Feb 25 10:39:00.778: INFO: Deleting pod "simpletest.rc-tlcnr" in namespace "gc-2058" +Feb 25 10:39:00.914: INFO: Deleting pod "simpletest.rc-v6qlc" in namespace "gc-2058" +Feb 25 10:39:01.364: INFO: Deleting pod "simpletest.rc-vxlkj" in namespace "gc-2058" +Feb 25 10:39:01.660: INFO: Deleting pod "simpletest.rc-w25hq" in namespace "gc-2058" +Feb 25 10:39:02.059: INFO: Deleting pod "simpletest.rc-w4lvz" in namespace "gc-2058" +Feb 25 10:39:02.175: INFO: Deleting pod "simpletest.rc-w646b" in namespace "gc-2058" +Feb 25 10:39:02.314: INFO: Deleting pod "simpletest.rc-w8hbl" in namespace "gc-2058" +Feb 25 10:39:02.382: INFO: Deleting pod "simpletest.rc-wd4mm" in namespace "gc-2058" +Feb 25 10:39:02.538: INFO: Deleting pod "simpletest.rc-wfd42" in namespace "gc-2058" +Feb 25 10:39:02.688: INFO: Deleting pod "simpletest.rc-whzt8" in namespace "gc-2058" +Feb 25 10:39:02.836: INFO: Deleting pod "simpletest.rc-xglb7" in namespace "gc-2058" +Feb 25 10:39:02.966: INFO: Deleting pod "simpletest.rc-xh6dc" in namespace "gc-2058" +Feb 25 10:39:03.059: INFO: Deleting pod "simpletest.rc-z9j8n" in namespace "gc-2058" +Feb 25 10:39:03.163: INFO: Deleting pod "simpletest.rc-zzmk9" in namespace "gc-2058" +[AfterEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:187 -Dec 14 16:34:00.220: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-7046" for this suite. 12/14/22 16:34:00.227 -{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","completed":174,"skipped":3139,"failed":0} +Feb 25 10:39:03.350: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "gc-2058" for this suite. 02/25/23 10:39:03.401 +{"msg":"PASSED [sig-api-machinery] Garbage collector should orphan pods created by rc if delete options say so [Conformance]","completed":171,"skipped":3241,"failed":0} ------------------------------ -• [4.235 seconds] -[sig-storage] Secrets -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:67 +• [SLOW TEST] [61.063 seconds] +[sig-api-machinery] Garbage collector +test/e2e/apimachinery/framework.go:23 + should orphan pods created by rc if delete options say so [Conformance] + test/e2e/apimachinery/garbage_collector.go:370 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Secrets + [BeforeEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:33:56.008 - Dec 14 16:33:56.008: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename secrets 12/14/22 16:33:56.031 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:33:56.107 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:33:56.116 - [It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:67 - STEP: Creating secret with name secret-test-f694f753-3e75-4e66-8fb5-e7e9243f5200 12/14/22 16:33:56.12 - STEP: Creating a pod to test consume secrets 12/14/22 16:33:56.135 - Dec 14 16:33:56.149: INFO: Waiting up to 5m0s for pod "pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3" in namespace "secrets-7046" to be "Succeeded or Failed" - Dec 14 16:33:56.163: INFO: Pod "pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3": Phase="Pending", Reason="", readiness=false. Elapsed: 13.642254ms - Dec 14 16:33:58.173: INFO: Pod "pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023490789s - Dec 14 16:34:00.170: INFO: Pod "pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021094448s - STEP: Saw pod success 12/14/22 16:34:00.17 - Dec 14 16:34:00.171: INFO: Pod "pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3" satisfied condition "Succeeded or Failed" - Dec 14 16:34:00.175: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3 container secret-volume-test: - STEP: delete the pod 12/14/22 16:34:00.185 - Dec 14 16:34:00.212: INFO: Waiting for pod pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3 to disappear - Dec 14 16:34:00.219: INFO: Pod pod-secrets-da72f90f-5829-42aa-85fc-d0d8140b16d3 no longer exists - [AfterEach] [sig-storage] Secrets - test/e2e/framework/framework.go:187 - Dec 14 16:34:00.220: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "secrets-7046" for this suite. 12/14/22 16:34:00.227 - << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSS ------------------------------- -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - should mutate custom resource with pruning [Conformance] - test/e2e/apimachinery/webhook.go:340 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:34:00.246 -Dec 14 16:34:00.246: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 16:34:00.257 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:00.287 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:00.291 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 16:34:00.312 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:34:01.375 -STEP: Deploying the webhook pod 12/14/22 16:34:01.384 -STEP: Wait for the deployment to be ready 12/14/22 16:34:01.402 -Dec 14 16:34:01.412: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created -STEP: Deploying the webhook service 12/14/22 16:34:03.428 -STEP: Verifying the service has paired with the endpoint 12/14/22 16:34:03.443 -Dec 14 16:34:04.443: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] should mutate custom resource with pruning [Conformance] - test/e2e/apimachinery/webhook.go:340 -Dec 14 16:34:04.450: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Registering the mutating webhook for custom resource e2e-test-webhook-2774-crds.webhook.example.com via the AdmissionRegistration API 12/14/22 16:34:04.972 -STEP: Creating a custom resource that should be mutated by the webhook 12/14/22 16:34:05.017 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/framework/framework.go:187 -Dec 14 16:34:07.749: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-2854" for this suite. 12/14/22 16:34:07.76 -STEP: Destroying namespace "webhook-2854-markers" for this suite. 12/14/22 16:34:07.785 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with pruning [Conformance]","completed":175,"skipped":3147,"failed":0} ------------------------------- -• [SLOW TEST] [7.715 seconds] -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - should mutate custom resource with pruning [Conformance] - test/e2e/apimachinery/webhook.go:340 + STEP: Creating a kubernetes client 02/25/23 10:38:02.418 + Feb 25 10:38:02.418: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename gc 02/25/23 10:38:02.424 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:38:02.469 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:38:02.476 + [It] should orphan pods created by rc if delete options say so [Conformance] + test/e2e/apimachinery/garbage_collector.go:370 + STEP: create the rc 02/25/23 10:38:02.492 + STEP: delete the rc 02/25/23 10:38:07.519 + STEP: wait for the rc to be deleted 02/25/23 10:38:07.675 + STEP: wait for 30 seconds to see if the garbage collector mistakenly deletes the pods 02/25/23 10:38:12.694 + STEP: Gathering metrics 02/25/23 10:38:43.149 + Feb 25 10:38:43.710: INFO: Waiting up to 5m0s for pod "kube-controller-manager-diek3dua9gay-2" in namespace "kube-system" to be "running and ready" + Feb 25 10:38:43.717: INFO: Pod "kube-controller-manager-diek3dua9gay-2": Phase="Running", Reason="", readiness=true. Elapsed: 6.760534ms + Feb 25 10:38:43.717: INFO: The phase of Pod kube-controller-manager-diek3dua9gay-2 is Running (Ready = true) + Feb 25 10:38:43.717: INFO: Pod "kube-controller-manager-diek3dua9gay-2" satisfied condition "running and ready" + Feb 25 10:38:43.886: INFO: For apiserver_request_total: + For apiserver_request_latency_seconds: + For apiserver_init_events_total: + For garbage_collector_attempt_to_delete_queue_latency: + For garbage_collector_attempt_to_delete_work_duration: + For garbage_collector_attempt_to_orphan_queue_latency: + For garbage_collector_attempt_to_orphan_work_duration: + For garbage_collector_dirty_processing_latency_microseconds: + For garbage_collector_event_processing_latency_microseconds: + For garbage_collector_graph_changes_queue_latency: + For garbage_collector_graph_changes_work_duration: + For garbage_collector_orphan_processing_latency_microseconds: + For namespace_queue_latency: + For namespace_queue_latency_sum: + For namespace_queue_latency_count: + For namespace_retries: + For namespace_work_duration: + For namespace_work_duration_sum: + For namespace_work_duration_count: + For function_duration_seconds: + For errors_total: + For evicted_pods_total: - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:34:00.246 - Dec 14 16:34:00.246: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 16:34:00.257 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:00.287 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:00.291 - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 16:34:00.312 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:34:01.375 - STEP: Deploying the webhook pod 12/14/22 16:34:01.384 - STEP: Wait for the deployment to be ready 12/14/22 16:34:01.402 - Dec 14 16:34:01.412: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created - STEP: Deploying the webhook service 12/14/22 16:34:03.428 - STEP: Verifying the service has paired with the endpoint 12/14/22 16:34:03.443 - Dec 14 16:34:04.443: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] should mutate custom resource with pruning [Conformance] - test/e2e/apimachinery/webhook.go:340 - Dec 14 16:34:04.450: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Registering the mutating webhook for custom resource e2e-test-webhook-2774-crds.webhook.example.com via the AdmissionRegistration API 12/14/22 16:34:04.972 - STEP: Creating a custom resource that should be mutated by the webhook 12/14/22 16:34:05.017 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + Feb 25 10:38:43.903: INFO: Deleting pod "simpletest.rc-2jbch" in namespace "gc-2058" + Feb 25 10:38:44.055: INFO: Deleting pod "simpletest.rc-2jjdn" in namespace "gc-2058" + Feb 25 10:38:44.206: INFO: Deleting pod "simpletest.rc-2vw46" in namespace "gc-2058" + Feb 25 10:38:44.347: INFO: Deleting pod "simpletest.rc-2zcch" in namespace "gc-2058" + Feb 25 10:38:44.422: INFO: Deleting pod "simpletest.rc-4hcl9" in namespace "gc-2058" + Feb 25 10:38:44.519: INFO: Deleting pod "simpletest.rc-4lvpz" in namespace "gc-2058" + Feb 25 10:38:44.783: INFO: Deleting pod "simpletest.rc-5467n" in namespace "gc-2058" + Feb 25 10:38:45.774: INFO: Deleting pod "simpletest.rc-5s9xb" in namespace "gc-2058" + Feb 25 10:38:46.400: INFO: Deleting pod "simpletest.rc-66vp6" in namespace "gc-2058" + Feb 25 10:38:46.502: INFO: Deleting pod "simpletest.rc-6j6df" in namespace "gc-2058" + Feb 25 10:38:46.594: INFO: Deleting pod "simpletest.rc-6kd5m" in namespace "gc-2058" + Feb 25 10:38:46.661: INFO: Deleting pod "simpletest.rc-6s54n" in namespace "gc-2058" + Feb 25 10:38:46.776: INFO: Deleting pod "simpletest.rc-6w59b" in namespace "gc-2058" + Feb 25 10:38:47.084: INFO: Deleting pod "simpletest.rc-7ffrx" in namespace "gc-2058" + Feb 25 10:38:47.338: INFO: Deleting pod "simpletest.rc-7h6vq" in namespace "gc-2058" + Feb 25 10:38:47.393: INFO: Deleting pod "simpletest.rc-7pr6f" in namespace "gc-2058" + Feb 25 10:38:47.466: INFO: Deleting pod "simpletest.rc-7pwx5" in namespace "gc-2058" + Feb 25 10:38:47.625: INFO: Deleting pod "simpletest.rc-7rrw4" in namespace "gc-2058" + Feb 25 10:38:47.786: INFO: Deleting pod "simpletest.rc-87pr4" in namespace "gc-2058" + Feb 25 10:38:47.901: INFO: Deleting pod "simpletest.rc-8pvw4" in namespace "gc-2058" + Feb 25 10:38:47.991: INFO: Deleting pod "simpletest.rc-8rs56" in namespace "gc-2058" + Feb 25 10:38:48.099: INFO: Deleting pod "simpletest.rc-962sn" in namespace "gc-2058" + Feb 25 10:38:48.177: INFO: Deleting pod "simpletest.rc-98982" in namespace "gc-2058" + Feb 25 10:38:48.252: INFO: Deleting pod "simpletest.rc-99xpb" in namespace "gc-2058" + Feb 25 10:38:48.426: INFO: Deleting pod "simpletest.rc-9pc9r" in namespace "gc-2058" + Feb 25 10:38:48.631: INFO: Deleting pod "simpletest.rc-b5qsf" in namespace "gc-2058" + Feb 25 10:38:48.732: INFO: Deleting pod "simpletest.rc-b687j" in namespace "gc-2058" + Feb 25 10:38:48.870: INFO: Deleting pod "simpletest.rc-bcxts" in namespace "gc-2058" + Feb 25 10:38:49.475: INFO: Deleting pod "simpletest.rc-bhnwx" in namespace "gc-2058" + Feb 25 10:38:49.866: INFO: Deleting pod "simpletest.rc-bt82t" in namespace "gc-2058" + Feb 25 10:38:50.001: INFO: Deleting pod "simpletest.rc-cbzzk" in namespace "gc-2058" + Feb 25 10:38:50.144: INFO: Deleting pod "simpletest.rc-cnkzc" in namespace "gc-2058" + Feb 25 10:38:50.310: INFO: Deleting pod "simpletest.rc-cxmjx" in namespace "gc-2058" + Feb 25 10:38:50.467: INFO: Deleting pod "simpletest.rc-d6gcj" in namespace "gc-2058" + Feb 25 10:38:50.579: INFO: Deleting pod "simpletest.rc-d8pft" in namespace "gc-2058" + Feb 25 10:38:50.671: INFO: Deleting pod "simpletest.rc-dnrb9" in namespace "gc-2058" + Feb 25 10:38:50.836: INFO: Deleting pod "simpletest.rc-drjpz" in namespace "gc-2058" + Feb 25 10:38:50.977: INFO: Deleting pod "simpletest.rc-f57nc" in namespace "gc-2058" + Feb 25 10:38:51.100: INFO: Deleting pod "simpletest.rc-f66j9" in namespace "gc-2058" + Feb 25 10:38:51.213: INFO: Deleting pod "simpletest.rc-f9cc9" in namespace "gc-2058" + Feb 25 10:38:51.277: INFO: Deleting pod "simpletest.rc-f9rwm" in namespace "gc-2058" + Feb 25 10:38:51.434: INFO: Deleting pod "simpletest.rc-ffsbr" in namespace "gc-2058" + Feb 25 10:38:51.544: INFO: Deleting pod "simpletest.rc-fqhc2" in namespace "gc-2058" + Feb 25 10:38:51.673: INFO: Deleting pod "simpletest.rc-g4td6" in namespace "gc-2058" + Feb 25 10:38:51.841: INFO: Deleting pod "simpletest.rc-gfwgm" in namespace "gc-2058" + Feb 25 10:38:51.971: INFO: Deleting pod "simpletest.rc-ghb54" in namespace "gc-2058" + Feb 25 10:38:52.084: INFO: Deleting pod "simpletest.rc-gtzxp" in namespace "gc-2058" + Feb 25 10:38:52.244: INFO: Deleting pod "simpletest.rc-hnxpc" in namespace "gc-2058" + Feb 25 10:38:52.322: INFO: Deleting pod "simpletest.rc-j2b7m" in namespace "gc-2058" + Feb 25 10:38:52.480: INFO: Deleting pod "simpletest.rc-j565f" in namespace "gc-2058" + Feb 25 10:38:52.628: INFO: Deleting pod "simpletest.rc-j5gsp" in namespace "gc-2058" + Feb 25 10:38:52.763: INFO: Deleting pod "simpletest.rc-j9r2p" in namespace "gc-2058" + Feb 25 10:38:53.177: INFO: Deleting pod "simpletest.rc-jd4jq" in namespace "gc-2058" + Feb 25 10:38:54.007: INFO: Deleting pod "simpletest.rc-jm7wv" in namespace "gc-2058" + Feb 25 10:38:54.298: INFO: Deleting pod "simpletest.rc-jvk6q" in namespace "gc-2058" + Feb 25 10:38:54.635: INFO: Deleting pod "simpletest.rc-k7gtr" in namespace "gc-2058" + Feb 25 10:38:54.864: INFO: Deleting pod "simpletest.rc-klxbd" in namespace "gc-2058" + Feb 25 10:38:55.053: INFO: Deleting pod "simpletest.rc-kvdtj" in namespace "gc-2058" + Feb 25 10:38:55.286: INFO: Deleting pod "simpletest.rc-l2pks" in namespace "gc-2058" + Feb 25 10:38:55.358: INFO: Deleting pod "simpletest.rc-lg9fl" in namespace "gc-2058" + Feb 25 10:38:55.456: INFO: Deleting pod "simpletest.rc-lrmzl" in namespace "gc-2058" + Feb 25 10:38:55.654: INFO: Deleting pod "simpletest.rc-lxtjn" in namespace "gc-2058" + Feb 25 10:38:55.981: INFO: Deleting pod "simpletest.rc-md8kp" in namespace "gc-2058" + Feb 25 10:38:56.137: INFO: Deleting pod "simpletest.rc-mgshb" in namespace "gc-2058" + Feb 25 10:38:56.275: INFO: Deleting pod "simpletest.rc-mjbj5" in namespace "gc-2058" + Feb 25 10:38:56.427: INFO: Deleting pod "simpletest.rc-n228r" in namespace "gc-2058" + Feb 25 10:38:56.587: INFO: Deleting pod "simpletest.rc-n5np6" in namespace "gc-2058" + Feb 25 10:38:56.700: INFO: Deleting pod "simpletest.rc-nhxxc" in namespace "gc-2058" + Feb 25 10:38:56.823: INFO: Deleting pod "simpletest.rc-njr4p" in namespace "gc-2058" + Feb 25 10:38:57.101: INFO: Deleting pod "simpletest.rc-nx4nd" in namespace "gc-2058" + Feb 25 10:38:58.178: INFO: Deleting pod "simpletest.rc-pntg2" in namespace "gc-2058" + Feb 25 10:38:58.381: INFO: Deleting pod "simpletest.rc-prq7h" in namespace "gc-2058" + Feb 25 10:38:58.698: INFO: Deleting pod "simpletest.rc-q99c2" in namespace "gc-2058" + Feb 25 10:38:58.807: INFO: Deleting pod "simpletest.rc-qq4cx" in namespace "gc-2058" + Feb 25 10:38:58.878: INFO: Deleting pod "simpletest.rc-qtwtp" in namespace "gc-2058" + Feb 25 10:38:59.127: INFO: Deleting pod "simpletest.rc-r55lj" in namespace "gc-2058" + Feb 25 10:38:59.286: INFO: Deleting pod "simpletest.rc-rc9kb" in namespace "gc-2058" + Feb 25 10:38:59.457: INFO: Deleting pod "simpletest.rc-rjw55" in namespace "gc-2058" + Feb 25 10:38:59.720: INFO: Deleting pod "simpletest.rc-rrq5p" in namespace "gc-2058" + Feb 25 10:38:59.870: INFO: Deleting pod "simpletest.rc-s5gth" in namespace "gc-2058" + Feb 25 10:39:00.041: INFO: Deleting pod "simpletest.rc-scqdw" in namespace "gc-2058" + Feb 25 10:39:00.108: INFO: Deleting pod "simpletest.rc-sl5fr" in namespace "gc-2058" + Feb 25 10:39:00.328: INFO: Deleting pod "simpletest.rc-srtb2" in namespace "gc-2058" + Feb 25 10:39:00.427: INFO: Deleting pod "simpletest.rc-ssv48" in namespace "gc-2058" + Feb 25 10:39:00.562: INFO: Deleting pod "simpletest.rc-swjdw" in namespace "gc-2058" + Feb 25 10:39:00.666: INFO: Deleting pod "simpletest.rc-tcth5" in namespace "gc-2058" + Feb 25 10:39:00.778: INFO: Deleting pod "simpletest.rc-tlcnr" in namespace "gc-2058" + Feb 25 10:39:00.914: INFO: Deleting pod "simpletest.rc-v6qlc" in namespace "gc-2058" + Feb 25 10:39:01.364: INFO: Deleting pod "simpletest.rc-vxlkj" in namespace "gc-2058" + Feb 25 10:39:01.660: INFO: Deleting pod "simpletest.rc-w25hq" in namespace "gc-2058" + Feb 25 10:39:02.059: INFO: Deleting pod "simpletest.rc-w4lvz" in namespace "gc-2058" + Feb 25 10:39:02.175: INFO: Deleting pod "simpletest.rc-w646b" in namespace "gc-2058" + Feb 25 10:39:02.314: INFO: Deleting pod "simpletest.rc-w8hbl" in namespace "gc-2058" + Feb 25 10:39:02.382: INFO: Deleting pod "simpletest.rc-wd4mm" in namespace "gc-2058" + Feb 25 10:39:02.538: INFO: Deleting pod "simpletest.rc-wfd42" in namespace "gc-2058" + Feb 25 10:39:02.688: INFO: Deleting pod "simpletest.rc-whzt8" in namespace "gc-2058" + Feb 25 10:39:02.836: INFO: Deleting pod "simpletest.rc-xglb7" in namespace "gc-2058" + Feb 25 10:39:02.966: INFO: Deleting pod "simpletest.rc-xh6dc" in namespace "gc-2058" + Feb 25 10:39:03.059: INFO: Deleting pod "simpletest.rc-z9j8n" in namespace "gc-2058" + Feb 25 10:39:03.163: INFO: Deleting pod "simpletest.rc-zzmk9" in namespace "gc-2058" + [AfterEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:187 - Dec 14 16:34:07.749: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-2854" for this suite. 12/14/22 16:34:07.76 - STEP: Destroying namespace "webhook-2854-markers" for this suite. 12/14/22 16:34:07.785 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 + Feb 25 10:39:03.350: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "gc-2058" for this suite. 02/25/23 10:39:03.401 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-cli] Kubectl client Kubectl cluster-info - should check if Kubernetes control plane services is included in cluster-info [Conformance] - test/e2e/kubectl/kubectl.go:1248 -[BeforeEach] [sig-cli] Kubectl client +[sig-node] RuntimeClass + should reject a Pod requesting a non-existent RuntimeClass [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:55 +[BeforeEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:34:07.975 -Dec 14 16:34:07.975: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 16:34:07.977 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:08.025 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:08.028 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 -[It] should check if Kubernetes control plane services is included in cluster-info [Conformance] - test/e2e/kubectl/kubectl.go:1248 -STEP: validating cluster-info 12/14/22 16:34:08.032 -Dec 14 16:34:08.032: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2269 cluster-info' -Dec 14 16:34:08.229: INFO: stderr: "" -Dec 14 16:34:08.230: INFO: stdout: "\x1b[0;32mKubernetes control plane\x1b[0m is running at \x1b[0;33mhttps://10.233.0.1:443\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n" -[AfterEach] [sig-cli] Kubectl client +STEP: Creating a kubernetes client 02/25/23 10:39:03.517 +Feb 25 10:39:03.518: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename runtimeclass 02/25/23 10:39:03.535 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:39:03.62 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:39:03.628 +[It] should reject a Pod requesting a non-existent RuntimeClass [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:55 +[AfterEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:187 -Dec 14 16:34:08.231: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-2269" for this suite. 12/14/22 16:34:08.238 -{"msg":"PASSED [sig-cli] Kubectl client Kubectl cluster-info should check if Kubernetes control plane services is included in cluster-info [Conformance]","completed":176,"skipped":3219,"failed":0} +Feb 25 10:39:03.672: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "runtimeclass-5159" for this suite. 02/25/23 10:39:03.707 +{"msg":"PASSED [sig-node] RuntimeClass should reject a Pod requesting a non-existent RuntimeClass [NodeConformance] [Conformance]","completed":172,"skipped":3281,"failed":0} ------------------------------ -• [0.272 seconds] -[sig-cli] Kubectl client -test/e2e/kubectl/framework.go:23 - Kubectl cluster-info - test/e2e/kubectl/kubectl.go:1242 - should check if Kubernetes control plane services is included in cluster-info [Conformance] - test/e2e/kubectl/kubectl.go:1248 +• [0.231 seconds] +[sig-node] RuntimeClass +test/e2e/common/node/framework.go:23 + should reject a Pod requesting a non-existent RuntimeClass [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:55 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-cli] Kubectl client + [BeforeEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:34:07.975 - Dec 14 16:34:07.975: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 16:34:07.977 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:08.025 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:08.028 - [BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 - [It] should check if Kubernetes control plane services is included in cluster-info [Conformance] - test/e2e/kubectl/kubectl.go:1248 - STEP: validating cluster-info 12/14/22 16:34:08.032 - Dec 14 16:34:08.032: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2269 cluster-info' - Dec 14 16:34:08.229: INFO: stderr: "" - Dec 14 16:34:08.230: INFO: stdout: "\x1b[0;32mKubernetes control plane\x1b[0m is running at \x1b[0;33mhttps://10.233.0.1:443\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n" - [AfterEach] [sig-cli] Kubectl client + STEP: Creating a kubernetes client 02/25/23 10:39:03.517 + Feb 25 10:39:03.518: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename runtimeclass 02/25/23 10:39:03.535 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:39:03.62 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:39:03.628 + [It] should reject a Pod requesting a non-existent RuntimeClass [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:55 + [AfterEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:187 - Dec 14 16:34:08.231: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-2269" for this suite. 12/14/22 16:34:08.238 + Feb 25 10:39:03.672: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "runtimeclass-5159" for this suite. 02/25/23 10:39:03.707 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-apps] Job - should create pods for an Indexed job with completion indexes and specified hostname [Conformance] - test/e2e/apps/job.go:194 -[BeforeEach] [sig-apps] Job +[sig-storage] Projected configMap + should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:374 +[BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:34:08.248 -Dec 14 16:34:08.248: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename job 12/14/22 16:34:08.252 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:08.29 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:08.299 -[It] should create pods for an Indexed job with completion indexes and specified hostname [Conformance] - test/e2e/apps/job.go:194 -STEP: Creating Indexed job 12/14/22 16:34:08.304 -STEP: Ensuring job reaches completions 12/14/22 16:34:08.317 -STEP: Ensuring pods with index for job exist 12/14/22 16:34:18.325 -[AfterEach] [sig-apps] Job +STEP: Creating a kubernetes client 02/25/23 10:39:03.758 +Feb 25 10:39:03.759: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 10:39:03.765 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:39:03.871 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:39:03.88 +[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:374 +STEP: Creating configMap with name projected-configmap-test-volume-4617e482-b068-49e8-9dc1-16cc2170a293 02/25/23 10:39:03.887 +STEP: Creating a pod to test consume configMaps 02/25/23 10:39:03.947 +Feb 25 10:39:03.979: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a" in namespace "projected-4597" to be "Succeeded or Failed" +Feb 25 10:39:03.996: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 16.954015ms +Feb 25 10:39:06.028: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.048631212s +Feb 25 10:39:08.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.026525488s +Feb 25 10:39:10.013: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 6.03322527s +Feb 25 10:39:12.008: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 8.028469165s +Feb 25 10:39:14.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 10.026444914s +Feb 25 10:39:16.004: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 12.024744001s +Feb 25 10:39:18.005: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 14.025191424s +Feb 25 10:39:20.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 16.027067288s +Feb 25 10:39:22.008: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 18.028869391s +Feb 25 10:39:24.005: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 20.025458474s +Feb 25 10:39:26.004: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 22.025108593s +Feb 25 10:39:28.005: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 24.025390954s +Feb 25 10:39:30.007: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 26.02764419s +Feb 25 10:39:32.003: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 28.023779506s +Feb 25 10:39:34.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 30.026561031s +Feb 25 10:39:36.004: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 32.024974758s +Feb 25 10:39:38.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 34.026693723s +Feb 25 10:39:40.007: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 36.02781358s +Feb 25 10:39:42.008: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 38.028413976s +Feb 25 10:39:44.007: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 40.02718344s +Feb 25 10:39:46.005: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 42.025932214s +Feb 25 10:39:48.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 44.026285607s +Feb 25 10:39:50.004: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 46.024636479s +Feb 25 10:39:52.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 48.026727449s +Feb 25 10:39:54.007: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 50.027597988s +Feb 25 10:39:56.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 52.026226675s +Feb 25 10:39:58.012: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 54.032386134s +Feb 25 10:40:00.010: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 56.030396522s +Feb 25 10:40:02.007: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 58.028052019s +STEP: Saw pod success 02/25/23 10:40:02.008 +Feb 25 10:40:02.008: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a" satisfied condition "Succeeded or Failed" +Feb 25 10:40:02.019: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a container projected-configmap-volume-test: +STEP: delete the pod 02/25/23 10:40:02.207 +Feb 25 10:40:02.242: INFO: Waiting for pod pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a to disappear +Feb 25 10:40:02.249: INFO: Pod pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a no longer exists +[AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 -Dec 14 16:34:18.332: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "job-4795" for this suite. 12/14/22 16:34:18.339 -{"msg":"PASSED [sig-apps] Job should create pods for an Indexed job with completion indexes and specified hostname [Conformance]","completed":177,"skipped":3226,"failed":0} +Feb 25 10:40:02.250: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-4597" for this suite. 02/25/23 10:40:02.26 +{"msg":"PASSED [sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","completed":173,"skipped":3359,"failed":0} ------------------------------ -• [SLOW TEST] [10.099 seconds] -[sig-apps] Job -test/e2e/apps/framework.go:23 - should create pods for an Indexed job with completion indexes and specified hostname [Conformance] - test/e2e/apps/job.go:194 +• [SLOW TEST] [58.519 seconds] +[sig-storage] Projected configMap +test/e2e/common/storage/framework.go:23 + should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:374 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Job + [BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:34:08.248 - Dec 14 16:34:08.248: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename job 12/14/22 16:34:08.252 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:08.29 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:08.299 - [It] should create pods for an Indexed job with completion indexes and specified hostname [Conformance] - test/e2e/apps/job.go:194 - STEP: Creating Indexed job 12/14/22 16:34:08.304 - STEP: Ensuring job reaches completions 12/14/22 16:34:08.317 - STEP: Ensuring pods with index for job exist 12/14/22 16:34:18.325 - [AfterEach] [sig-apps] Job + STEP: Creating a kubernetes client 02/25/23 10:39:03.758 + Feb 25 10:39:03.759: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 10:39:03.765 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:39:03.871 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:39:03.88 + [It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:374 + STEP: Creating configMap with name projected-configmap-test-volume-4617e482-b068-49e8-9dc1-16cc2170a293 02/25/23 10:39:03.887 + STEP: Creating a pod to test consume configMaps 02/25/23 10:39:03.947 + Feb 25 10:39:03.979: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a" in namespace "projected-4597" to be "Succeeded or Failed" + Feb 25 10:39:03.996: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 16.954015ms + Feb 25 10:39:06.028: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.048631212s + Feb 25 10:39:08.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.026525488s + Feb 25 10:39:10.013: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 6.03322527s + Feb 25 10:39:12.008: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 8.028469165s + Feb 25 10:39:14.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 10.026444914s + Feb 25 10:39:16.004: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 12.024744001s + Feb 25 10:39:18.005: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 14.025191424s + Feb 25 10:39:20.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 16.027067288s + Feb 25 10:39:22.008: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 18.028869391s + Feb 25 10:39:24.005: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 20.025458474s + Feb 25 10:39:26.004: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 22.025108593s + Feb 25 10:39:28.005: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 24.025390954s + Feb 25 10:39:30.007: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 26.02764419s + Feb 25 10:39:32.003: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 28.023779506s + Feb 25 10:39:34.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 30.026561031s + Feb 25 10:39:36.004: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 32.024974758s + Feb 25 10:39:38.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 34.026693723s + Feb 25 10:39:40.007: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 36.02781358s + Feb 25 10:39:42.008: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 38.028413976s + Feb 25 10:39:44.007: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 40.02718344s + Feb 25 10:39:46.005: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 42.025932214s + Feb 25 10:39:48.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 44.026285607s + Feb 25 10:39:50.004: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 46.024636479s + Feb 25 10:39:52.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 48.026727449s + Feb 25 10:39:54.007: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 50.027597988s + Feb 25 10:39:56.006: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 52.026226675s + Feb 25 10:39:58.012: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 54.032386134s + Feb 25 10:40:00.010: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Pending", Reason="", readiness=false. Elapsed: 56.030396522s + Feb 25 10:40:02.007: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 58.028052019s + STEP: Saw pod success 02/25/23 10:40:02.008 + Feb 25 10:40:02.008: INFO: Pod "pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a" satisfied condition "Succeeded or Failed" + Feb 25 10:40:02.019: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a container projected-configmap-volume-test: + STEP: delete the pod 02/25/23 10:40:02.207 + Feb 25 10:40:02.242: INFO: Waiting for pod pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a to disappear + Feb 25 10:40:02.249: INFO: Pod pod-projected-configmaps-53a09717-4161-4d74-8960-d2ac72fda73a no longer exists + [AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 - Dec 14 16:34:18.332: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "job-4795" for this suite. 12/14/22 16:34:18.339 + Feb 25 10:40:02.250: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-4597" for this suite. 02/25/23 10:40:02.26 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSS +SSSSSSS ------------------------------ -[sig-apps] Deployment - should validate Deployment Status endpoints [Conformance] - test/e2e/apps/deployment.go:479 -[BeforeEach] [sig-apps] Deployment +[sig-api-machinery] Garbage collector + should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] + test/e2e/apimachinery/garbage_collector.go:650 +[BeforeEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:34:18.354 -Dec 14 16:34:18.355: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename deployment 12/14/22 16:34:18.356 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:18.383 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:18.386 -[BeforeEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:91 -[It] should validate Deployment Status endpoints [Conformance] - test/e2e/apps/deployment.go:479 -STEP: creating a Deployment 12/14/22 16:34:18.395 -Dec 14 16:34:18.395: INFO: Creating simple deployment test-deployment-w7dwv -Dec 14 16:34:18.416: INFO: new replicaset for deployment "test-deployment-w7dwv" is yet to be created -STEP: Getting /status 12/14/22 16:34:20.459 -Dec 14 16:34:20.476: INFO: Deployment test-deployment-w7dwv has Conditions: [{Available True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:19 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-w7dwv-777898ffcc" has successfully progressed.}] -STEP: updating Deployment Status 12/14/22 16:34:20.476 -Dec 14 16:34:20.500: INFO: updatedStatus.Conditions: []v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 34, 19, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 34, 19, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 34, 19, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 34, 18, 0, time.Local), Reason:"NewReplicaSetAvailable", Message:"ReplicaSet \"test-deployment-w7dwv-777898ffcc\" has successfully progressed."}, v1.DeploymentCondition{Type:"StatusUpdate", Status:"True", LastUpdateTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} -STEP: watching for the Deployment status to be updated 12/14/22 16:34:20.5 -Dec 14 16:34:20.503: INFO: Observed &Deployment event: ADDED -Dec 14 16:34:20.503: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-w7dwv-777898ffcc"} -Dec 14 16:34:20.503: INFO: Observed &Deployment event: MODIFIED -Dec 14 16:34:20.503: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-w7dwv-777898ffcc"} -Dec 14 16:34:20.503: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} -Dec 14 16:34:20.504: INFO: Observed &Deployment event: MODIFIED -Dec 14 16:34:20.504: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} -Dec 14 16:34:20.504: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC ReplicaSetUpdated ReplicaSet "test-deployment-w7dwv-777898ffcc" is progressing.} -Dec 14 16:34:20.504: INFO: Observed &Deployment event: MODIFIED -Dec 14 16:34:20.504: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:19 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} -Dec 14 16:34:20.504: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-w7dwv-777898ffcc" has successfully progressed.} -Dec 14 16:34:20.505: INFO: Observed &Deployment event: MODIFIED -Dec 14 16:34:20.505: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:19 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} -Dec 14 16:34:20.505: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-w7dwv-777898ffcc" has successfully progressed.} -Dec 14 16:34:20.505: INFO: Found Deployment test-deployment-w7dwv in namespace deployment-9032 with labels: map[e2e:testing name:httpd] annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} -Dec 14 16:34:20.505: INFO: Deployment test-deployment-w7dwv has an updated status -STEP: patching the Statefulset Status 12/14/22 16:34:20.505 -Dec 14 16:34:20.507: INFO: Patch payload: {"status":{"conditions":[{"type":"StatusPatched","status":"True"}]}} -Dec 14 16:34:20.515: INFO: Patched status conditions: []v1.DeploymentCondition{v1.DeploymentCondition{Type:"StatusPatched", Status:"True", LastUpdateTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"", Message:""}} -STEP: watching for the Deployment status to be patched 12/14/22 16:34:20.515 -Dec 14 16:34:20.518: INFO: Observed &Deployment event: ADDED -Dec 14 16:34:20.518: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-w7dwv-777898ffcc"} -Dec 14 16:34:20.518: INFO: Observed &Deployment event: MODIFIED -Dec 14 16:34:20.518: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-w7dwv-777898ffcc"} -Dec 14 16:34:20.518: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} -Dec 14 16:34:20.519: INFO: Observed &Deployment event: MODIFIED -Dec 14 16:34:20.519: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} -Dec 14 16:34:20.519: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC ReplicaSetUpdated ReplicaSet "test-deployment-w7dwv-777898ffcc" is progressing.} -Dec 14 16:34:20.519: INFO: Observed &Deployment event: MODIFIED -Dec 14 16:34:20.519: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:19 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} -Dec 14 16:34:20.519: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-w7dwv-777898ffcc" has successfully progressed.} -Dec 14 16:34:20.520: INFO: Observed &Deployment event: MODIFIED -Dec 14 16:34:20.520: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:19 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} -Dec 14 16:34:20.520: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-w7dwv-777898ffcc" has successfully progressed.} -Dec 14 16:34:20.520: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} -Dec 14 16:34:20.520: INFO: Observed &Deployment event: MODIFIED -Dec 14 16:34:20.520: INFO: Found deployment test-deployment-w7dwv in namespace deployment-9032 with labels: map[e2e:testing name:httpd] annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {StatusPatched True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } -Dec 14 16:34:20.520: INFO: Deployment test-deployment-w7dwv has a patched status -[AfterEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:84 -Dec 14 16:34:20.525: INFO: Deployment "test-deployment-w7dwv": -&Deployment{ObjectMeta:{test-deployment-w7dwv deployment-9032 755886e7-e080-4492-9cb3-c17b1482d418 19705 1 2022-12-14 16:34:18 +0000 UTC map[e2e:testing name:httpd] map[deployment.kubernetes.io/revision:1] [] [] [{e2e.test Update apps/v1 2022-12-14 16:34:18 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:e2e":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:e2e":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {e2e.test Update apps/v1 2022-12-14 16:34:20 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"StatusPatched\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:status":{},"f:type":{}}}}} status} {kube-controller-manager Update apps/v1 2022-12-14 16:34:20 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{e2e: testing,name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[e2e:testing name:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0004019a8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:StatusPatched,Status:True,Reason:,Message:,LastUpdateTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:0001-01-01 00:00:00 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:FoundNewReplicaSet,Message:Found new replica set "test-deployment-w7dwv-777898ffcc",LastUpdateTime:2022-12-14 16:34:20 +0000 UTC,LastTransitionTime:2022-12-14 16:34:20 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} +STEP: Creating a kubernetes client 02/25/23 10:40:02.29 +Feb 25 10:40:02.290: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename gc 02/25/23 10:40:02.296 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:40:02.339 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:40:02.346 +[It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] + test/e2e/apimachinery/garbage_collector.go:650 +STEP: create the rc 02/25/23 10:40:02.364 +STEP: delete the rc 02/25/23 10:40:12.83 +STEP: wait for the rc to be deleted 02/25/23 10:40:12.999 +Feb 25 10:40:14.296: INFO: 100 pods remaining +Feb 25 10:40:14.296: INFO: 98 pods has nil DeletionTimestamp +Feb 25 10:40:14.296: INFO: +Feb 25 10:40:15.722: INFO: 91 pods remaining +Feb 25 10:40:15.722: INFO: 81 pods has nil DeletionTimestamp +Feb 25 10:40:15.722: INFO: +Feb 25 10:40:16.793: INFO: 82 pods remaining +Feb 25 10:40:16.793: INFO: 71 pods has nil DeletionTimestamp +Feb 25 10:40:16.793: INFO: +Feb 25 10:40:18.063: INFO: 75 pods remaining +Feb 25 10:40:18.064: INFO: 60 pods has nil DeletionTimestamp +Feb 25 10:40:18.064: INFO: +Feb 25 10:40:20.972: INFO: 62 pods remaining +Feb 25 10:40:20.972: INFO: 46 pods has nil DeletionTimestamp +Feb 25 10:40:20.972: INFO: +Feb 25 10:40:23.669: INFO: 54 pods remaining +Feb 25 10:40:23.670: INFO: 34 pods has nil DeletionTimestamp +Feb 25 10:40:23.670: INFO: +Feb 25 10:40:25.485: INFO: 45 pods remaining +Feb 25 10:40:25.486: INFO: 18 pods has nil DeletionTimestamp +Feb 25 10:40:25.486: INFO: +Feb 25 10:40:27.513: INFO: 36 pods remaining +Feb 25 10:40:27.513: INFO: 6 pods has nil DeletionTimestamp +Feb 25 10:40:27.513: INFO: +Feb 25 10:40:28.262: INFO: 34 pods remaining +Feb 25 10:40:28.263: INFO: 1 pods has nil DeletionTimestamp +Feb 25 10:40:28.263: INFO: +Feb 25 10:40:29.321: INFO: 24 pods remaining +Feb 25 10:40:29.321: INFO: 0 pods has nil DeletionTimestamp +Feb 25 10:40:29.321: INFO: +Feb 25 10:40:31.328: INFO: 11 pods remaining +Feb 25 10:40:31.328: INFO: 0 pods has nil DeletionTimestamp +Feb 25 10:40:31.328: INFO: +Feb 25 10:40:32.120: INFO: 1 pods remaining +Feb 25 10:40:32.121: INFO: 0 pods has nil DeletionTimestamp +Feb 25 10:40:32.121: INFO: +STEP: Gathering metrics 02/25/23 10:40:33.516 +Feb 25 10:40:34.321: INFO: Waiting up to 5m0s for pod "kube-controller-manager-diek3dua9gay-2" in namespace "kube-system" to be "running and ready" +Feb 25 10:40:34.587: INFO: Pod "kube-controller-manager-diek3dua9gay-2": Phase="Running", Reason="", readiness=true. Elapsed: 265.599762ms +Feb 25 10:40:34.587: INFO: The phase of Pod kube-controller-manager-diek3dua9gay-2 is Running (Ready = true) +Feb 25 10:40:34.587: INFO: Pod "kube-controller-manager-diek3dua9gay-2" satisfied condition "running and ready" +Feb 25 10:40:36.968: INFO: For apiserver_request_total: +For apiserver_request_latency_seconds: +For apiserver_init_events_total: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: -Dec 14 16:34:20.528: INFO: New ReplicaSet "test-deployment-w7dwv-777898ffcc" of Deployment "test-deployment-w7dwv": -&ReplicaSet{ObjectMeta:{test-deployment-w7dwv-777898ffcc deployment-9032 f56c9bb5-21af-4bd6-88df-a4c0750a236c 19698 1 2022-12-14 16:34:18 +0000 UTC map[e2e:testing name:httpd pod-template-hash:777898ffcc] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-deployment-w7dwv 755886e7-e080-4492-9cb3-c17b1482d418 0xc001e22840 0xc001e22841}] [] [{kube-controller-manager Update apps/v1 2022-12-14 16:34:18 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:e2e":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"755886e7-e080-4492-9cb3-c17b1482d418\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:e2e":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 16:34:19 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{e2e: testing,name: httpd,pod-template-hash: 777898ffcc,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[e2e:testing name:httpd pod-template-hash:777898ffcc] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc001e228e8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} -Dec 14 16:34:20.533: INFO: Pod "test-deployment-w7dwv-777898ffcc-nz8fk" is available: -&Pod{ObjectMeta:{test-deployment-w7dwv-777898ffcc-nz8fk test-deployment-w7dwv-777898ffcc- deployment-9032 f03d6662-ef42-46aa-90f1-0bd115364291 19697 0 2022-12-14 16:34:18 +0000 UTC map[e2e:testing name:httpd pod-template-hash:777898ffcc] map[] [{apps/v1 ReplicaSet test-deployment-w7dwv-777898ffcc f56c9bb5-21af-4bd6-88df-a4c0750a236c 0xc001e22ca0 0xc001e22ca1}] [] [{kube-controller-manager Update v1 2022-12-14 16:34:18 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:e2e":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f56c9bb5-21af-4bd6-88df-a4c0750a236c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:34:19 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.10\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-x29jb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x29jb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:34:18 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:34:19 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:34:19 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:34:18 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.10,StartTime:2022-12-14 16:34:18 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:34:18 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://c1bd58a4e2017e653eb49165dd8b3377cfa8367103134714339296fa4b007a6d,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.10,},},EphemeralContainerStatuses:[]ContainerStatus{},},} -[AfterEach] [sig-apps] Deployment +[AfterEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:187 -Dec 14 16:34:20.533: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "deployment-9032" for this suite. 12/14/22 16:34:20.541 -{"msg":"PASSED [sig-apps] Deployment should validate Deployment Status endpoints [Conformance]","completed":178,"skipped":3245,"failed":0} +Feb 25 10:40:36.969: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "gc-6951" for this suite. 02/25/23 10:40:37.003 +{"msg":"PASSED [sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]","completed":174,"skipped":3366,"failed":0} ------------------------------ -• [2.198 seconds] -[sig-apps] Deployment -test/e2e/apps/framework.go:23 - should validate Deployment Status endpoints [Conformance] - test/e2e/apps/deployment.go:479 +• [SLOW TEST] [34.898 seconds] +[sig-api-machinery] Garbage collector +test/e2e/apimachinery/framework.go:23 + should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] + test/e2e/apimachinery/garbage_collector.go:650 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Deployment + [BeforeEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:34:18.354 - Dec 14 16:34:18.355: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename deployment 12/14/22 16:34:18.356 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:18.383 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:18.386 - [BeforeEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:91 - [It] should validate Deployment Status endpoints [Conformance] - test/e2e/apps/deployment.go:479 - STEP: creating a Deployment 12/14/22 16:34:18.395 - Dec 14 16:34:18.395: INFO: Creating simple deployment test-deployment-w7dwv - Dec 14 16:34:18.416: INFO: new replicaset for deployment "test-deployment-w7dwv" is yet to be created - STEP: Getting /status 12/14/22 16:34:20.459 - Dec 14 16:34:20.476: INFO: Deployment test-deployment-w7dwv has Conditions: [{Available True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:19 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} {Progressing True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-w7dwv-777898ffcc" has successfully progressed.}] - STEP: updating Deployment Status 12/14/22 16:34:20.476 - Dec 14 16:34:20.500: INFO: updatedStatus.Conditions: []v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 34, 19, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 34, 19, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 34, 19, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 34, 18, 0, time.Local), Reason:"NewReplicaSetAvailable", Message:"ReplicaSet \"test-deployment-w7dwv-777898ffcc\" has successfully progressed."}, v1.DeploymentCondition{Type:"StatusUpdate", Status:"True", LastUpdateTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} - STEP: watching for the Deployment status to be updated 12/14/22 16:34:20.5 - Dec 14 16:34:20.503: INFO: Observed &Deployment event: ADDED - Dec 14 16:34:20.503: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-w7dwv-777898ffcc"} - Dec 14 16:34:20.503: INFO: Observed &Deployment event: MODIFIED - Dec 14 16:34:20.503: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-w7dwv-777898ffcc"} - Dec 14 16:34:20.503: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} - Dec 14 16:34:20.504: INFO: Observed &Deployment event: MODIFIED - Dec 14 16:34:20.504: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} - Dec 14 16:34:20.504: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC ReplicaSetUpdated ReplicaSet "test-deployment-w7dwv-777898ffcc" is progressing.} - Dec 14 16:34:20.504: INFO: Observed &Deployment event: MODIFIED - Dec 14 16:34:20.504: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:19 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} - Dec 14 16:34:20.504: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-w7dwv-777898ffcc" has successfully progressed.} - Dec 14 16:34:20.505: INFO: Observed &Deployment event: MODIFIED - Dec 14 16:34:20.505: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:19 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} - Dec 14 16:34:20.505: INFO: Observed Deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-w7dwv-777898ffcc" has successfully progressed.} - Dec 14 16:34:20.505: INFO: Found Deployment test-deployment-w7dwv in namespace deployment-9032 with labels: map[e2e:testing name:httpd] annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} - Dec 14 16:34:20.505: INFO: Deployment test-deployment-w7dwv has an updated status - STEP: patching the Statefulset Status 12/14/22 16:34:20.505 - Dec 14 16:34:20.507: INFO: Patch payload: {"status":{"conditions":[{"type":"StatusPatched","status":"True"}]}} - Dec 14 16:34:20.515: INFO: Patched status conditions: []v1.DeploymentCondition{v1.DeploymentCondition{Type:"StatusPatched", Status:"True", LastUpdateTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"", Message:""}} - STEP: watching for the Deployment status to be patched 12/14/22 16:34:20.515 - Dec 14 16:34:20.518: INFO: Observed &Deployment event: ADDED - Dec 14 16:34:20.518: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-w7dwv-777898ffcc"} - Dec 14 16:34:20.518: INFO: Observed &Deployment event: MODIFIED - Dec 14 16:34:20.518: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetCreated Created new replica set "test-deployment-w7dwv-777898ffcc"} - Dec 14 16:34:20.518: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} - Dec 14 16:34:20.519: INFO: Observed &Deployment event: MODIFIED - Dec 14 16:34:20.519: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available False 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC MinimumReplicasUnavailable Deployment does not have minimum availability.} - Dec 14 16:34:20.519: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:18 +0000 UTC 2022-12-14 16:34:18 +0000 UTC ReplicaSetUpdated ReplicaSet "test-deployment-w7dwv-777898ffcc" is progressing.} - Dec 14 16:34:20.519: INFO: Observed &Deployment event: MODIFIED - Dec 14 16:34:20.519: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:19 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} - Dec 14 16:34:20.519: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-w7dwv-777898ffcc" has successfully progressed.} - Dec 14 16:34:20.520: INFO: Observed &Deployment event: MODIFIED - Dec 14 16:34:20.520: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Available True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:19 +0000 UTC MinimumReplicasAvailable Deployment has minimum availability.} - Dec 14 16:34:20.520: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {Progressing True 2022-12-14 16:34:19 +0000 UTC 2022-12-14 16:34:18 +0000 UTC NewReplicaSetAvailable ReplicaSet "test-deployment-w7dwv-777898ffcc" has successfully progressed.} - Dec 14 16:34:20.520: INFO: Observed deployment test-deployment-w7dwv in namespace deployment-9032 with annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {StatusUpdate True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test} - Dec 14 16:34:20.520: INFO: Observed &Deployment event: MODIFIED - Dec 14 16:34:20.520: INFO: Found deployment test-deployment-w7dwv in namespace deployment-9032 with labels: map[e2e:testing name:httpd] annotations: map[deployment.kubernetes.io/revision:1] & Conditions: {StatusPatched True 0001-01-01 00:00:00 +0000 UTC 0001-01-01 00:00:00 +0000 UTC } - Dec 14 16:34:20.520: INFO: Deployment test-deployment-w7dwv has a patched status - [AfterEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:84 - Dec 14 16:34:20.525: INFO: Deployment "test-deployment-w7dwv": - &Deployment{ObjectMeta:{test-deployment-w7dwv deployment-9032 755886e7-e080-4492-9cb3-c17b1482d418 19705 1 2022-12-14 16:34:18 +0000 UTC map[e2e:testing name:httpd] map[deployment.kubernetes.io/revision:1] [] [] [{e2e.test Update apps/v1 2022-12-14 16:34:18 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:e2e":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:e2e":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {e2e.test Update apps/v1 2022-12-14 16:34:20 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"StatusPatched\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:status":{},"f:type":{}}}}} status} {kube-controller-manager Update apps/v1 2022-12-14 16:34:20 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{e2e: testing,name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[e2e:testing name:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0004019a8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:StatusPatched,Status:True,Reason:,Message:,LastUpdateTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:0001-01-01 00:00:00 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:FoundNewReplicaSet,Message:Found new replica set "test-deployment-w7dwv-777898ffcc",LastUpdateTime:2022-12-14 16:34:20 +0000 UTC,LastTransitionTime:2022-12-14 16:34:20 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} + STEP: Creating a kubernetes client 02/25/23 10:40:02.29 + Feb 25 10:40:02.290: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename gc 02/25/23 10:40:02.296 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:40:02.339 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:40:02.346 + [It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] + test/e2e/apimachinery/garbage_collector.go:650 + STEP: create the rc 02/25/23 10:40:02.364 + STEP: delete the rc 02/25/23 10:40:12.83 + STEP: wait for the rc to be deleted 02/25/23 10:40:12.999 + Feb 25 10:40:14.296: INFO: 100 pods remaining + Feb 25 10:40:14.296: INFO: 98 pods has nil DeletionTimestamp + Feb 25 10:40:14.296: INFO: + Feb 25 10:40:15.722: INFO: 91 pods remaining + Feb 25 10:40:15.722: INFO: 81 pods has nil DeletionTimestamp + Feb 25 10:40:15.722: INFO: + Feb 25 10:40:16.793: INFO: 82 pods remaining + Feb 25 10:40:16.793: INFO: 71 pods has nil DeletionTimestamp + Feb 25 10:40:16.793: INFO: + Feb 25 10:40:18.063: INFO: 75 pods remaining + Feb 25 10:40:18.064: INFO: 60 pods has nil DeletionTimestamp + Feb 25 10:40:18.064: INFO: + Feb 25 10:40:20.972: INFO: 62 pods remaining + Feb 25 10:40:20.972: INFO: 46 pods has nil DeletionTimestamp + Feb 25 10:40:20.972: INFO: + Feb 25 10:40:23.669: INFO: 54 pods remaining + Feb 25 10:40:23.670: INFO: 34 pods has nil DeletionTimestamp + Feb 25 10:40:23.670: INFO: + Feb 25 10:40:25.485: INFO: 45 pods remaining + Feb 25 10:40:25.486: INFO: 18 pods has nil DeletionTimestamp + Feb 25 10:40:25.486: INFO: + Feb 25 10:40:27.513: INFO: 36 pods remaining + Feb 25 10:40:27.513: INFO: 6 pods has nil DeletionTimestamp + Feb 25 10:40:27.513: INFO: + Feb 25 10:40:28.262: INFO: 34 pods remaining + Feb 25 10:40:28.263: INFO: 1 pods has nil DeletionTimestamp + Feb 25 10:40:28.263: INFO: + Feb 25 10:40:29.321: INFO: 24 pods remaining + Feb 25 10:40:29.321: INFO: 0 pods has nil DeletionTimestamp + Feb 25 10:40:29.321: INFO: + Feb 25 10:40:31.328: INFO: 11 pods remaining + Feb 25 10:40:31.328: INFO: 0 pods has nil DeletionTimestamp + Feb 25 10:40:31.328: INFO: + Feb 25 10:40:32.120: INFO: 1 pods remaining + Feb 25 10:40:32.121: INFO: 0 pods has nil DeletionTimestamp + Feb 25 10:40:32.121: INFO: + STEP: Gathering metrics 02/25/23 10:40:33.516 + Feb 25 10:40:34.321: INFO: Waiting up to 5m0s for pod "kube-controller-manager-diek3dua9gay-2" in namespace "kube-system" to be "running and ready" + Feb 25 10:40:34.587: INFO: Pod "kube-controller-manager-diek3dua9gay-2": Phase="Running", Reason="", readiness=true. Elapsed: 265.599762ms + Feb 25 10:40:34.587: INFO: The phase of Pod kube-controller-manager-diek3dua9gay-2 is Running (Ready = true) + Feb 25 10:40:34.587: INFO: Pod "kube-controller-manager-diek3dua9gay-2" satisfied condition "running and ready" + Feb 25 10:40:36.968: INFO: For apiserver_request_total: + For apiserver_request_latency_seconds: + For apiserver_init_events_total: + For garbage_collector_attempt_to_delete_queue_latency: + For garbage_collector_attempt_to_delete_work_duration: + For garbage_collector_attempt_to_orphan_queue_latency: + For garbage_collector_attempt_to_orphan_work_duration: + For garbage_collector_dirty_processing_latency_microseconds: + For garbage_collector_event_processing_latency_microseconds: + For garbage_collector_graph_changes_queue_latency: + For garbage_collector_graph_changes_work_duration: + For garbage_collector_orphan_processing_latency_microseconds: + For namespace_queue_latency: + For namespace_queue_latency_sum: + For namespace_queue_latency_count: + For namespace_retries: + For namespace_work_duration: + For namespace_work_duration_sum: + For namespace_work_duration_count: + For function_duration_seconds: + For errors_total: + For evicted_pods_total: - Dec 14 16:34:20.528: INFO: New ReplicaSet "test-deployment-w7dwv-777898ffcc" of Deployment "test-deployment-w7dwv": - &ReplicaSet{ObjectMeta:{test-deployment-w7dwv-777898ffcc deployment-9032 f56c9bb5-21af-4bd6-88df-a4c0750a236c 19698 1 2022-12-14 16:34:18 +0000 UTC map[e2e:testing name:httpd pod-template-hash:777898ffcc] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-deployment-w7dwv 755886e7-e080-4492-9cb3-c17b1482d418 0xc001e22840 0xc001e22841}] [] [{kube-controller-manager Update apps/v1 2022-12-14 16:34:18 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:e2e":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"755886e7-e080-4492-9cb3-c17b1482d418\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:e2e":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 16:34:19 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{e2e: testing,name: httpd,pod-template-hash: 777898ffcc,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[e2e:testing name:httpd pod-template-hash:777898ffcc] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc001e228e8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} - Dec 14 16:34:20.533: INFO: Pod "test-deployment-w7dwv-777898ffcc-nz8fk" is available: - &Pod{ObjectMeta:{test-deployment-w7dwv-777898ffcc-nz8fk test-deployment-w7dwv-777898ffcc- deployment-9032 f03d6662-ef42-46aa-90f1-0bd115364291 19697 0 2022-12-14 16:34:18 +0000 UTC map[e2e:testing name:httpd pod-template-hash:777898ffcc] map[] [{apps/v1 ReplicaSet test-deployment-w7dwv-777898ffcc f56c9bb5-21af-4bd6-88df-a4c0750a236c 0xc001e22ca0 0xc001e22ca1}] [] [{kube-controller-manager Update v1 2022-12-14 16:34:18 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:e2e":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f56c9bb5-21af-4bd6-88df-a4c0750a236c\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:34:19 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.10\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-x29jb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-x29jb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:34:18 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:34:19 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:34:19 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:34:18 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.10,StartTime:2022-12-14 16:34:18 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:34:18 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://c1bd58a4e2017e653eb49165dd8b3377cfa8367103134714339296fa4b007a6d,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.10,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - [AfterEach] [sig-apps] Deployment + [AfterEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:187 - Dec 14 16:34:20.533: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "deployment-9032" for this suite. 12/14/22 16:34:20.541 + Feb 25 10:40:36.969: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "gc-6951" for this suite. 02/25/23 10:40:37.003 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSS ------------------------------ -[sig-node] Pods - should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:397 -[BeforeEach] [sig-node] Pods +[sig-storage] EmptyDir volumes + should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:126 +[BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:34:20.56 -Dec 14 16:34:20.560: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename pods 12/14/22 16:34:20.567 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:20.598 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:20.602 -[BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 -[It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:397 -STEP: creating the pod 12/14/22 16:34:20.606 -STEP: submitting the pod to kubernetes 12/14/22 16:34:20.607 -Dec 14 16:34:20.617: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330" in namespace "pods-8041" to be "running and ready" -Dec 14 16:34:20.621: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330": Phase="Pending", Reason="", readiness=false. Elapsed: 3.997747ms -Dec 14 16:34:20.621: INFO: The phase of Pod pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:34:22.629: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330": Phase="Running", Reason="", readiness=true. Elapsed: 2.012127513s -Dec 14 16:34:22.629: INFO: The phase of Pod pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330 is Running (Ready = true) -Dec 14 16:34:22.629: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330" satisfied condition "running and ready" -STEP: verifying the pod is in kubernetes 12/14/22 16:34:22.634 -STEP: updating the pod 12/14/22 16:34:22.64 -Dec 14 16:34:23.154: INFO: Successfully updated pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330" -Dec 14 16:34:23.154: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330" in namespace "pods-8041" to be "terminated with reason DeadlineExceeded" -Dec 14 16:34:23.170: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330": Phase="Running", Reason="", readiness=true. Elapsed: 15.439472ms -Dec 14 16:34:25.178: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330": Phase="Running", Reason="", readiness=true. Elapsed: 2.023498339s -Dec 14 16:34:27.179: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330": Phase="Running", Reason="", readiness=false. Elapsed: 4.024809532s -Dec 14 16:34:29.178: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 6.02390038s -Dec 14 16:34:29.178: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330" satisfied condition "terminated with reason DeadlineExceeded" -[AfterEach] [sig-node] Pods +STEP: Creating a kubernetes client 02/25/23 10:40:37.188 +Feb 25 10:40:37.188: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir 02/25/23 10:40:37.197 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:40:38.02 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:40:38.046 +[It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:126 +STEP: Creating a pod to test emptydir 0644 on tmpfs 02/25/23 10:40:38.055 +Feb 25 10:40:38.106: INFO: Waiting up to 5m0s for pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498" in namespace "emptydir-7529" to be "Succeeded or Failed" +Feb 25 10:40:38.121: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 14.87741ms +Feb 25 10:40:40.163: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 2.057162036s +Feb 25 10:40:42.177: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 4.071318155s +Feb 25 10:40:44.214: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 6.107453187s +Feb 25 10:40:46.141: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 8.035189129s +Feb 25 10:40:48.149: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 10.042637895s +Feb 25 10:40:50.130: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 12.024247125s +Feb 25 10:40:52.148: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 14.041652863s +Feb 25 10:40:54.131: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 16.025172442s +Feb 25 10:40:56.135: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 18.028626263s +Feb 25 10:40:58.137: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 20.030539426s +Feb 25 10:41:00.137: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 22.030804976s +Feb 25 10:41:02.165: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 24.059147173s +Feb 25 10:41:04.143: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 26.036975284s +Feb 25 10:41:06.138: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 28.031779037s +Feb 25 10:41:08.137: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 30.030649342s +Feb 25 10:41:10.134: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 32.028073283s +Feb 25 10:41:12.137: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 34.030514192s +Feb 25 10:41:14.132: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 36.025917795s +Feb 25 10:41:16.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 38.026592514s +Feb 25 10:41:18.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 40.026823152s +Feb 25 10:41:20.135: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 42.029051241s +Feb 25 10:41:22.135: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 44.028491778s +Feb 25 10:41:24.138: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 46.031949956s +Feb 25 10:41:26.144: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 48.037885534s +Feb 25 10:41:28.130: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 50.0236771s +Feb 25 10:41:30.132: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 52.02555083s +Feb 25 10:41:32.140: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 54.034174021s +Feb 25 10:41:34.130: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 56.02360618s +Feb 25 10:41:36.132: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 58.025650404s +Feb 25 10:41:38.130: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m0.023926115s +Feb 25 10:41:40.128: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m2.022364477s +Feb 25 10:41:42.134: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m4.027761213s +Feb 25 10:41:44.140: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m6.033455396s +Feb 25 10:41:46.130: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m8.023797297s +Feb 25 10:41:48.134: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m10.027402096s +Feb 25 10:41:50.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m12.027206027s +Feb 25 10:41:52.141: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m14.034559455s +Feb 25 10:41:54.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m16.026916482s +Feb 25 10:41:56.143: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m18.036522628s +Feb 25 10:41:58.144: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m20.037828716s +Feb 25 10:42:00.129: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m22.023010307s +Feb 25 10:42:02.145: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m24.038784553s +Feb 25 10:42:04.135: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m26.028769067s +Feb 25 10:42:06.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m28.02677572s +Feb 25 10:42:08.131: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m30.025035076s +Feb 25 10:42:10.131: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m32.02491884s +Feb 25 10:42:12.130: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m34.02408102s +Feb 25 10:42:14.135: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m36.028876068s +Feb 25 10:42:16.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m38.02637454s +Feb 25 10:42:18.132: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m40.025761075s +Feb 25 10:42:20.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m42.026733931s +Feb 25 10:42:22.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m44.026827591s +Feb 25 10:42:24.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m46.026653659s +Feb 25 10:42:26.135: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m48.028641301s +Feb 25 10:42:28.132: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m50.025557503s +Feb 25 10:42:30.131: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m52.024862026s +Feb 25 10:42:32.134: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Succeeded", Reason="", readiness=false. Elapsed: 1m54.027584691s +STEP: Saw pod success 02/25/23 10:42:32.134 +Feb 25 10:42:32.134: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498" satisfied condition "Succeeded or Failed" +Feb 25 10:42:32.148: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-3dab1caf-867d-4f94-ab21-625bb7767498 container test-container: +STEP: delete the pod 02/25/23 10:42:32.248 +Feb 25 10:42:32.274: INFO: Waiting for pod pod-3dab1caf-867d-4f94-ab21-625bb7767498 to disappear +Feb 25 10:42:32.284: INFO: Pod pod-3dab1caf-867d-4f94-ab21-625bb7767498 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 -Dec 14 16:34:29.179: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-8041" for this suite. 12/14/22 16:34:29.185 -{"msg":"PASSED [sig-node] Pods should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance]","completed":179,"skipped":3325,"failed":0} +Feb 25 10:42:32.284: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-7529" for this suite. 02/25/23 10:42:32.295 +{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","completed":175,"skipped":3370,"failed":0} ------------------------------ -• [SLOW TEST] [8.634 seconds] -[sig-node] Pods -test/e2e/common/node/framework.go:23 - should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:397 +• [SLOW TEST] [115.124 seconds] +[sig-storage] EmptyDir volumes +test/e2e/common/storage/framework.go:23 + should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:126 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Pods + [BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:34:20.56 - Dec 14 16:34:20.560: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename pods 12/14/22 16:34:20.567 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:20.598 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:20.602 - [BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 - [It] should allow activeDeadlineSeconds to be updated [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:397 - STEP: creating the pod 12/14/22 16:34:20.606 - STEP: submitting the pod to kubernetes 12/14/22 16:34:20.607 - Dec 14 16:34:20.617: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330" in namespace "pods-8041" to be "running and ready" - Dec 14 16:34:20.621: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330": Phase="Pending", Reason="", readiness=false. Elapsed: 3.997747ms - Dec 14 16:34:20.621: INFO: The phase of Pod pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:34:22.629: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330": Phase="Running", Reason="", readiness=true. Elapsed: 2.012127513s - Dec 14 16:34:22.629: INFO: The phase of Pod pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330 is Running (Ready = true) - Dec 14 16:34:22.629: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330" satisfied condition "running and ready" - STEP: verifying the pod is in kubernetes 12/14/22 16:34:22.634 - STEP: updating the pod 12/14/22 16:34:22.64 - Dec 14 16:34:23.154: INFO: Successfully updated pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330" - Dec 14 16:34:23.154: INFO: Waiting up to 5m0s for pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330" in namespace "pods-8041" to be "terminated with reason DeadlineExceeded" - Dec 14 16:34:23.170: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330": Phase="Running", Reason="", readiness=true. Elapsed: 15.439472ms - Dec 14 16:34:25.178: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330": Phase="Running", Reason="", readiness=true. Elapsed: 2.023498339s - Dec 14 16:34:27.179: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330": Phase="Running", Reason="", readiness=false. Elapsed: 4.024809532s - Dec 14 16:34:29.178: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330": Phase="Failed", Reason="DeadlineExceeded", readiness=false. Elapsed: 6.02390038s - Dec 14 16:34:29.178: INFO: Pod "pod-update-activedeadlineseconds-52f6e0c2-7c87-47b8-8614-c0ee8e138330" satisfied condition "terminated with reason DeadlineExceeded" - [AfterEach] [sig-node] Pods + STEP: Creating a kubernetes client 02/25/23 10:40:37.188 + Feb 25 10:40:37.188: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir 02/25/23 10:40:37.197 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:40:38.02 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:40:38.046 + [It] should support (non-root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:126 + STEP: Creating a pod to test emptydir 0644 on tmpfs 02/25/23 10:40:38.055 + Feb 25 10:40:38.106: INFO: Waiting up to 5m0s for pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498" in namespace "emptydir-7529" to be "Succeeded or Failed" + Feb 25 10:40:38.121: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 14.87741ms + Feb 25 10:40:40.163: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 2.057162036s + Feb 25 10:40:42.177: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 4.071318155s + Feb 25 10:40:44.214: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 6.107453187s + Feb 25 10:40:46.141: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 8.035189129s + Feb 25 10:40:48.149: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 10.042637895s + Feb 25 10:40:50.130: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 12.024247125s + Feb 25 10:40:52.148: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 14.041652863s + Feb 25 10:40:54.131: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 16.025172442s + Feb 25 10:40:56.135: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 18.028626263s + Feb 25 10:40:58.137: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 20.030539426s + Feb 25 10:41:00.137: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 22.030804976s + Feb 25 10:41:02.165: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 24.059147173s + Feb 25 10:41:04.143: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 26.036975284s + Feb 25 10:41:06.138: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 28.031779037s + Feb 25 10:41:08.137: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 30.030649342s + Feb 25 10:41:10.134: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 32.028073283s + Feb 25 10:41:12.137: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 34.030514192s + Feb 25 10:41:14.132: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 36.025917795s + Feb 25 10:41:16.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 38.026592514s + Feb 25 10:41:18.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 40.026823152s + Feb 25 10:41:20.135: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 42.029051241s + Feb 25 10:41:22.135: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 44.028491778s + Feb 25 10:41:24.138: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 46.031949956s + Feb 25 10:41:26.144: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 48.037885534s + Feb 25 10:41:28.130: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 50.0236771s + Feb 25 10:41:30.132: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 52.02555083s + Feb 25 10:41:32.140: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 54.034174021s + Feb 25 10:41:34.130: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 56.02360618s + Feb 25 10:41:36.132: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 58.025650404s + Feb 25 10:41:38.130: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m0.023926115s + Feb 25 10:41:40.128: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m2.022364477s + Feb 25 10:41:42.134: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m4.027761213s + Feb 25 10:41:44.140: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m6.033455396s + Feb 25 10:41:46.130: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m8.023797297s + Feb 25 10:41:48.134: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m10.027402096s + Feb 25 10:41:50.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m12.027206027s + Feb 25 10:41:52.141: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m14.034559455s + Feb 25 10:41:54.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m16.026916482s + Feb 25 10:41:56.143: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m18.036522628s + Feb 25 10:41:58.144: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m20.037828716s + Feb 25 10:42:00.129: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m22.023010307s + Feb 25 10:42:02.145: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m24.038784553s + Feb 25 10:42:04.135: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m26.028769067s + Feb 25 10:42:06.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m28.02677572s + Feb 25 10:42:08.131: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m30.025035076s + Feb 25 10:42:10.131: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m32.02491884s + Feb 25 10:42:12.130: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m34.02408102s + Feb 25 10:42:14.135: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m36.028876068s + Feb 25 10:42:16.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m38.02637454s + Feb 25 10:42:18.132: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m40.025761075s + Feb 25 10:42:20.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m42.026733931s + Feb 25 10:42:22.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m44.026827591s + Feb 25 10:42:24.133: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m46.026653659s + Feb 25 10:42:26.135: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m48.028641301s + Feb 25 10:42:28.132: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m50.025557503s + Feb 25 10:42:30.131: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Pending", Reason="", readiness=false. Elapsed: 1m52.024862026s + Feb 25 10:42:32.134: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498": Phase="Succeeded", Reason="", readiness=false. Elapsed: 1m54.027584691s + STEP: Saw pod success 02/25/23 10:42:32.134 + Feb 25 10:42:32.134: INFO: Pod "pod-3dab1caf-867d-4f94-ab21-625bb7767498" satisfied condition "Succeeded or Failed" + Feb 25 10:42:32.148: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-3dab1caf-867d-4f94-ab21-625bb7767498 container test-container: + STEP: delete the pod 02/25/23 10:42:32.248 + Feb 25 10:42:32.274: INFO: Waiting for pod pod-3dab1caf-867d-4f94-ab21-625bb7767498 to disappear + Feb 25 10:42:32.284: INFO: Pod pod-3dab1caf-867d-4f94-ab21-625bb7767498 no longer exists + [AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 - Dec 14 16:34:29.179: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "pods-8041" for this suite. 12/14/22 16:34:29.185 + Feb 25 10:42:32.284: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-7529" for this suite. 02/25/23 10:42:32.295 << End Captured GinkgoWriter Output ------------------------------ -S +SSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] ConfigMap - should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:46 -[BeforeEach] [sig-storage] ConfigMap +[sig-network] DNS + should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] + test/e2e/network/dns.go:193 +[BeforeEach] [sig-network] DNS test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:34:29.196 -Dec 14 16:34:29.196: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename configmap 12/14/22 16:34:29.199 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:29.232 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:29.236 -[It] should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:46 -STEP: Creating configMap with name configmap-test-volume-39df1331-e8d3-45d6-928d-34194ddb467c 12/14/22 16:34:29.24 -STEP: Creating a pod to test consume configMaps 12/14/22 16:34:29.245 -Dec 14 16:34:29.257: INFO: Waiting up to 5m0s for pod "pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87" in namespace "configmap-8309" to be "Succeeded or Failed" -Dec 14 16:34:29.261: INFO: Pod "pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87": Phase="Pending", Reason="", readiness=false. Elapsed: 4.647865ms -Dec 14 16:34:31.268: INFO: Pod "pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011609318s -Dec 14 16:34:33.271: INFO: Pod "pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014701012s -STEP: Saw pod success 12/14/22 16:34:33.271 -Dec 14 16:34:33.272: INFO: Pod "pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87" satisfied condition "Succeeded or Failed" -Dec 14 16:34:33.276: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87 container agnhost-container: -STEP: delete the pod 12/14/22 16:34:33.287 -Dec 14 16:34:33.305: INFO: Waiting for pod pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87 to disappear -Dec 14 16:34:33.308: INFO: Pod pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87 no longer exists -[AfterEach] [sig-storage] ConfigMap +STEP: Creating a kubernetes client 02/25/23 10:42:32.343 +Feb 25 10:42:32.344: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename dns 02/25/23 10:42:32.352 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:42:32.401 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:42:32.406 +[It] should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] + test/e2e/network/dns.go:193 +STEP: Creating a test headless service 02/25/23 10:42:32.412 +STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6825 A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-6825;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6825 A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-6825;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6825.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-6825.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6825.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-6825.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-6825.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-6825.svc;check="$$(dig +notcp +noall +answer +search 208.46.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.46.208_udp@PTR;check="$$(dig +tcp +noall +answer +search 208.46.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.46.208_tcp@PTR;sleep 1; done + 02/25/23 10:42:32.466 +STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6825 A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-6825;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6825 A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-6825;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6825.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-6825.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6825.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-6825.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-6825.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-6825.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-6825.svc;check="$$(dig +notcp +noall +answer +search 208.46.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.46.208_udp@PTR;check="$$(dig +tcp +noall +answer +search 208.46.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.46.208_tcp@PTR;sleep 1; done + 02/25/23 10:42:32.467 +STEP: creating a pod to probe DNS 02/25/23 10:42:32.467 +STEP: submitting the pod to kubernetes 02/25/23 10:42:32.468 +Feb 25 10:42:32.506: INFO: Waiting up to 15m0s for pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069" in namespace "dns-6825" to be "running" +Feb 25 10:42:32.522: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 15.412894ms +Feb 25 10:42:34.537: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030978421s +Feb 25 10:42:36.543: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 4.037073795s +Feb 25 10:42:38.532: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 6.026343376s +Feb 25 10:42:40.537: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 8.030869989s +Feb 25 10:42:42.536: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 10.030263823s +Feb 25 10:42:44.535: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 12.028499493s +Feb 25 10:42:46.542: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 14.036355991s +Feb 25 10:42:48.532: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 16.025779037s +Feb 25 10:42:50.532: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 18.026301335s +Feb 25 10:42:52.535: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 20.028699208s +Feb 25 10:42:54.533: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 22.026481701s +Feb 25 10:42:56.531: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 24.024475633s +Feb 25 10:42:58.531: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 26.024863077s +Feb 25 10:43:00.532: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 28.025845745s +Feb 25 10:43:02.533: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 30.026445971s +Feb 25 10:43:04.536: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 32.02984683s +Feb 25 10:43:06.538: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 34.031613003s +Feb 25 10:43:08.535: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 36.028874769s +Feb 25 10:43:10.534: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 38.027823864s +Feb 25 10:43:12.531: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 40.025208725s +Feb 25 10:43:14.540: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 42.033555354s +Feb 25 10:43:16.530: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 44.023723479s +Feb 25 10:43:18.533: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 46.027218307s +Feb 25 10:43:20.537: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 48.030832938s +Feb 25 10:43:22.534: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 50.02806555s +Feb 25 10:43:24.533: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 52.027292097s +Feb 25 10:43:26.532: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 54.025815627s +Feb 25 10:43:28.531: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 56.025116668s +Feb 25 10:43:30.531: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 58.025213634s +Feb 25 10:43:32.532: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 1m0.025433286s +Feb 25 10:43:34.539: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 1m2.032448288s +Feb 25 10:43:36.532: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 1m4.025879507s +Feb 25 10:43:38.534: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 1m6.027693584s +Feb 25 10:43:40.535: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 1m8.029082789s +Feb 25 10:43:42.536: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Running", Reason="", readiness=true. Elapsed: 1m10.030172946s +Feb 25 10:43:42.536: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069" satisfied condition "running" +STEP: retrieving the pod 02/25/23 10:43:42.537 +STEP: looking for the results for each expected name from probers 02/25/23 10:43:42.561 +Feb 25 10:43:42.581: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.590: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.599: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.610: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.621: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.634: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.652: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.664: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.705: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.721: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.730: INFO: Unable to read jessie_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.738: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.752: INFO: Unable to read jessie_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.763: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.780: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.791: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:42.836: INFO: Lookups using dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6825 wheezy_tcp@dns-test-service.dns-6825 wheezy_udp@dns-test-service.dns-6825.svc wheezy_tcp@dns-test-service.dns-6825.svc wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6825 jessie_tcp@dns-test-service.dns-6825 jessie_udp@dns-test-service.dns-6825.svc jessie_tcp@dns-test-service.dns-6825.svc jessie_udp@_http._tcp.dns-test-service.dns-6825.svc jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc] + +Feb 25 10:43:47.860: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:47.903: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:47.918: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:47.928: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:47.941: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:47.954: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:47.964: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:47.978: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:48.058: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:48.072: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:48.082: INFO: Unable to read jessie_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:48.100: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:48.122: INFO: Unable to read jessie_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:48.136: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:48.149: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:48.168: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:48.227: INFO: Lookups using dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6825 wheezy_tcp@dns-test-service.dns-6825 wheezy_udp@dns-test-service.dns-6825.svc wheezy_tcp@dns-test-service.dns-6825.svc wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6825 jessie_tcp@dns-test-service.dns-6825 jessie_udp@dns-test-service.dns-6825.svc jessie_tcp@dns-test-service.dns-6825.svc jessie_udp@_http._tcp.dns-test-service.dns-6825.svc jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc] + +Feb 25 10:43:52.859: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:52.874: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:52.896: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:52.957: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:52.969: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:52.985: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:53.044: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:53.074: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:53.382: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:53.437: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:53.451: INFO: Unable to read jessie_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:53.504: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:53.518: INFO: Unable to read jessie_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:53.541: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:53.554: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:53.566: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:53.632: INFO: Lookups using dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6825 wheezy_tcp@dns-test-service.dns-6825 wheezy_udp@dns-test-service.dns-6825.svc wheezy_tcp@dns-test-service.dns-6825.svc wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6825 jessie_tcp@dns-test-service.dns-6825 jessie_udp@dns-test-service.dns-6825.svc jessie_tcp@dns-test-service.dns-6825.svc jessie_udp@_http._tcp.dns-test-service.dns-6825.svc jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc] + +Feb 25 10:43:57.851: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:57.872: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:57.904: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:57.912: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:57.936: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:57.969: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:57.981: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:57.992: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:58.078: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:58.098: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:58.106: INFO: Unable to read jessie_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:58.117: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:58.131: INFO: Unable to read jessie_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:58.143: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:58.154: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:58.174: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:43:58.247: INFO: Lookups using dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6825 wheezy_tcp@dns-test-service.dns-6825 wheezy_udp@dns-test-service.dns-6825.svc wheezy_tcp@dns-test-service.dns-6825.svc wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6825 jessie_tcp@dns-test-service.dns-6825 jessie_udp@dns-test-service.dns-6825.svc jessie_tcp@dns-test-service.dns-6825.svc jessie_udp@_http._tcp.dns-test-service.dns-6825.svc jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc] + +Feb 25 10:44:02.848: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:02.859: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:02.867: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:02.875: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:02.889: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:02.940: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:02.949: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:02.959: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:03.011: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:03.022: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:03.030: INFO: Unable to read jessie_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:03.041: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:03.056: INFO: Unable to read jessie_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:03.073: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:03.081: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:03.091: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:03.145: INFO: Lookups using dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6825 wheezy_tcp@dns-test-service.dns-6825 wheezy_udp@dns-test-service.dns-6825.svc wheezy_tcp@dns-test-service.dns-6825.svc wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6825 jessie_tcp@dns-test-service.dns-6825 jessie_udp@dns-test-service.dns-6825.svc jessie_tcp@dns-test-service.dns-6825.svc jessie_udp@_http._tcp.dns-test-service.dns-6825.svc jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc] + +Feb 25 10:44:07.860: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:07.869: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:07.876: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:07.886: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:07.898: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:07.905: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:07.916: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:07.935: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:07.976: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:07.984: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:07.990: INFO: Unable to read jessie_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:08.004: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:08.015: INFO: Unable to read jessie_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:08.021: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:08.030: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:08.038: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) +Feb 25 10:44:08.065: INFO: Lookups using dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6825 wheezy_tcp@dns-test-service.dns-6825 wheezy_udp@dns-test-service.dns-6825.svc wheezy_tcp@dns-test-service.dns-6825.svc wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6825 jessie_tcp@dns-test-service.dns-6825 jessie_udp@dns-test-service.dns-6825.svc jessie_tcp@dns-test-service.dns-6825.svc jessie_udp@_http._tcp.dns-test-service.dns-6825.svc jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc] + +Feb 25 10:44:13.681: INFO: DNS probes using dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069 succeeded + +STEP: deleting the pod 02/25/23 10:44:13.682 +STEP: deleting the test service 02/25/23 10:44:13.719 +STEP: deleting the test headless service 02/25/23 10:44:13.792 +[AfterEach] [sig-network] DNS test/e2e/framework/framework.go:187 -Dec 14 16:34:33.309: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-8309" for this suite. 12/14/22 16:34:33.315 -{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume [NodeConformance] [Conformance]","completed":180,"skipped":3326,"failed":0} +Feb 25 10:44:13.891: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "dns-6825" for this suite. 02/25/23 10:44:13.994 +{"msg":"PASSED [sig-network] DNS should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]","completed":176,"skipped":3387,"failed":0} ------------------------------ -• [4.131 seconds] -[sig-storage] ConfigMap -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:46 +• [SLOW TEST] [101.678 seconds] +[sig-network] DNS +test/e2e/network/common/framework.go:23 + should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] + test/e2e/network/dns.go:193 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] ConfigMap + [BeforeEach] [sig-network] DNS test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:34:29.196 - Dec 14 16:34:29.196: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename configmap 12/14/22 16:34:29.199 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:29.232 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:29.236 - [It] should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:46 - STEP: Creating configMap with name configmap-test-volume-39df1331-e8d3-45d6-928d-34194ddb467c 12/14/22 16:34:29.24 - STEP: Creating a pod to test consume configMaps 12/14/22 16:34:29.245 - Dec 14 16:34:29.257: INFO: Waiting up to 5m0s for pod "pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87" in namespace "configmap-8309" to be "Succeeded or Failed" - Dec 14 16:34:29.261: INFO: Pod "pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87": Phase="Pending", Reason="", readiness=false. Elapsed: 4.647865ms - Dec 14 16:34:31.268: INFO: Pod "pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011609318s - Dec 14 16:34:33.271: INFO: Pod "pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014701012s - STEP: Saw pod success 12/14/22 16:34:33.271 - Dec 14 16:34:33.272: INFO: Pod "pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87" satisfied condition "Succeeded or Failed" - Dec 14 16:34:33.276: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87 container agnhost-container: - STEP: delete the pod 12/14/22 16:34:33.287 - Dec 14 16:34:33.305: INFO: Waiting for pod pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87 to disappear - Dec 14 16:34:33.308: INFO: Pod pod-configmaps-f4ad5da0-5c20-40e9-9f12-4dca43b7ee87 no longer exists - [AfterEach] [sig-storage] ConfigMap + STEP: Creating a kubernetes client 02/25/23 10:42:32.343 + Feb 25 10:42:32.344: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename dns 02/25/23 10:42:32.352 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:42:32.401 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:42:32.406 + [It] should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] + test/e2e/network/dns.go:193 + STEP: Creating a test headless service 02/25/23 10:42:32.412 + STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6825 A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-6825;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6825 A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-6825;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6825.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-6825.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6825.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-6825.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-6825.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-6825.svc;check="$$(dig +notcp +noall +answer +search 208.46.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.46.208_udp@PTR;check="$$(dig +tcp +noall +answer +search 208.46.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.46.208_tcp@PTR;sleep 1; done + 02/25/23 10:42:32.466 + STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6825 A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-6825;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6825 A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-6825;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-6825.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-6825.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-6825.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-6825.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-6825.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-6825.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-6825.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-6825.svc;check="$$(dig +notcp +noall +answer +search 208.46.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.46.208_udp@PTR;check="$$(dig +tcp +noall +answer +search 208.46.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.46.208_tcp@PTR;sleep 1; done + 02/25/23 10:42:32.467 + STEP: creating a pod to probe DNS 02/25/23 10:42:32.467 + STEP: submitting the pod to kubernetes 02/25/23 10:42:32.468 + Feb 25 10:42:32.506: INFO: Waiting up to 15m0s for pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069" in namespace "dns-6825" to be "running" + Feb 25 10:42:32.522: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 15.412894ms + Feb 25 10:42:34.537: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030978421s + Feb 25 10:42:36.543: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 4.037073795s + Feb 25 10:42:38.532: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 6.026343376s + Feb 25 10:42:40.537: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 8.030869989s + Feb 25 10:42:42.536: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 10.030263823s + Feb 25 10:42:44.535: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 12.028499493s + Feb 25 10:42:46.542: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 14.036355991s + Feb 25 10:42:48.532: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 16.025779037s + Feb 25 10:42:50.532: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 18.026301335s + Feb 25 10:42:52.535: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 20.028699208s + Feb 25 10:42:54.533: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 22.026481701s + Feb 25 10:42:56.531: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 24.024475633s + Feb 25 10:42:58.531: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 26.024863077s + Feb 25 10:43:00.532: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 28.025845745s + Feb 25 10:43:02.533: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 30.026445971s + Feb 25 10:43:04.536: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 32.02984683s + Feb 25 10:43:06.538: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 34.031613003s + Feb 25 10:43:08.535: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 36.028874769s + Feb 25 10:43:10.534: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 38.027823864s + Feb 25 10:43:12.531: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 40.025208725s + Feb 25 10:43:14.540: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 42.033555354s + Feb 25 10:43:16.530: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 44.023723479s + Feb 25 10:43:18.533: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 46.027218307s + Feb 25 10:43:20.537: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 48.030832938s + Feb 25 10:43:22.534: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 50.02806555s + Feb 25 10:43:24.533: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 52.027292097s + Feb 25 10:43:26.532: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 54.025815627s + Feb 25 10:43:28.531: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 56.025116668s + Feb 25 10:43:30.531: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 58.025213634s + Feb 25 10:43:32.532: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 1m0.025433286s + Feb 25 10:43:34.539: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 1m2.032448288s + Feb 25 10:43:36.532: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 1m4.025879507s + Feb 25 10:43:38.534: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 1m6.027693584s + Feb 25 10:43:40.535: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Pending", Reason="", readiness=false. Elapsed: 1m8.029082789s + Feb 25 10:43:42.536: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069": Phase="Running", Reason="", readiness=true. Elapsed: 1m10.030172946s + Feb 25 10:43:42.536: INFO: Pod "dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069" satisfied condition "running" + STEP: retrieving the pod 02/25/23 10:43:42.537 + STEP: looking for the results for each expected name from probers 02/25/23 10:43:42.561 + Feb 25 10:43:42.581: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.590: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.599: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.610: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.621: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.634: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.652: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.664: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.705: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.721: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.730: INFO: Unable to read jessie_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.738: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.752: INFO: Unable to read jessie_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.763: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.780: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.791: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:42.836: INFO: Lookups using dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6825 wheezy_tcp@dns-test-service.dns-6825 wheezy_udp@dns-test-service.dns-6825.svc wheezy_tcp@dns-test-service.dns-6825.svc wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6825 jessie_tcp@dns-test-service.dns-6825 jessie_udp@dns-test-service.dns-6825.svc jessie_tcp@dns-test-service.dns-6825.svc jessie_udp@_http._tcp.dns-test-service.dns-6825.svc jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc] + + Feb 25 10:43:47.860: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:47.903: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:47.918: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:47.928: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:47.941: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:47.954: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:47.964: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:47.978: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:48.058: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:48.072: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:48.082: INFO: Unable to read jessie_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:48.100: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:48.122: INFO: Unable to read jessie_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:48.136: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:48.149: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:48.168: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:48.227: INFO: Lookups using dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6825 wheezy_tcp@dns-test-service.dns-6825 wheezy_udp@dns-test-service.dns-6825.svc wheezy_tcp@dns-test-service.dns-6825.svc wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6825 jessie_tcp@dns-test-service.dns-6825 jessie_udp@dns-test-service.dns-6825.svc jessie_tcp@dns-test-service.dns-6825.svc jessie_udp@_http._tcp.dns-test-service.dns-6825.svc jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc] + + Feb 25 10:43:52.859: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:52.874: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:52.896: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:52.957: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:52.969: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:52.985: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:53.044: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:53.074: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:53.382: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:53.437: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:53.451: INFO: Unable to read jessie_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:53.504: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:53.518: INFO: Unable to read jessie_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:53.541: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:53.554: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:53.566: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:53.632: INFO: Lookups using dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6825 wheezy_tcp@dns-test-service.dns-6825 wheezy_udp@dns-test-service.dns-6825.svc wheezy_tcp@dns-test-service.dns-6825.svc wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6825 jessie_tcp@dns-test-service.dns-6825 jessie_udp@dns-test-service.dns-6825.svc jessie_tcp@dns-test-service.dns-6825.svc jessie_udp@_http._tcp.dns-test-service.dns-6825.svc jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc] + + Feb 25 10:43:57.851: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:57.872: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:57.904: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:57.912: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:57.936: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:57.969: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:57.981: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:57.992: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:58.078: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:58.098: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:58.106: INFO: Unable to read jessie_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:58.117: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:58.131: INFO: Unable to read jessie_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:58.143: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:58.154: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:58.174: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:43:58.247: INFO: Lookups using dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6825 wheezy_tcp@dns-test-service.dns-6825 wheezy_udp@dns-test-service.dns-6825.svc wheezy_tcp@dns-test-service.dns-6825.svc wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6825 jessie_tcp@dns-test-service.dns-6825 jessie_udp@dns-test-service.dns-6825.svc jessie_tcp@dns-test-service.dns-6825.svc jessie_udp@_http._tcp.dns-test-service.dns-6825.svc jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc] + + Feb 25 10:44:02.848: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:02.859: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:02.867: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:02.875: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:02.889: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:02.940: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:02.949: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:02.959: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:03.011: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:03.022: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:03.030: INFO: Unable to read jessie_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:03.041: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:03.056: INFO: Unable to read jessie_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:03.073: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:03.081: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:03.091: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:03.145: INFO: Lookups using dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6825 wheezy_tcp@dns-test-service.dns-6825 wheezy_udp@dns-test-service.dns-6825.svc wheezy_tcp@dns-test-service.dns-6825.svc wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6825 jessie_tcp@dns-test-service.dns-6825 jessie_udp@dns-test-service.dns-6825.svc jessie_tcp@dns-test-service.dns-6825.svc jessie_udp@_http._tcp.dns-test-service.dns-6825.svc jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc] + + Feb 25 10:44:07.860: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:07.869: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:07.876: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:07.886: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:07.898: INFO: Unable to read wheezy_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:07.905: INFO: Unable to read wheezy_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:07.916: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:07.935: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:07.976: INFO: Unable to read jessie_udp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:07.984: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:07.990: INFO: Unable to read jessie_udp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:08.004: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825 from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:08.015: INFO: Unable to read jessie_udp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:08.021: INFO: Unable to read jessie_tcp@dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:08.030: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:08.038: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc from pod dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069: the server could not find the requested resource (get pods dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069) + Feb 25 10:44:08.065: INFO: Lookups using dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069 failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-6825 wheezy_tcp@dns-test-service.dns-6825 wheezy_udp@dns-test-service.dns-6825.svc wheezy_tcp@dns-test-service.dns-6825.svc wheezy_udp@_http._tcp.dns-test-service.dns-6825.svc wheezy_tcp@_http._tcp.dns-test-service.dns-6825.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-6825 jessie_tcp@dns-test-service.dns-6825 jessie_udp@dns-test-service.dns-6825.svc jessie_tcp@dns-test-service.dns-6825.svc jessie_udp@_http._tcp.dns-test-service.dns-6825.svc jessie_tcp@_http._tcp.dns-test-service.dns-6825.svc] + + Feb 25 10:44:13.681: INFO: DNS probes using dns-6825/dns-test-b3882823-473c-4b02-b3bb-7e85d4c01069 succeeded + + STEP: deleting the pod 02/25/23 10:44:13.682 + STEP: deleting the test service 02/25/23 10:44:13.719 + STEP: deleting the test headless service 02/25/23 10:44:13.792 + [AfterEach] [sig-network] DNS test/e2e/framework/framework.go:187 - Dec 14 16:34:33.309: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "configmap-8309" for this suite. 12/14/22 16:34:33.315 + Feb 25 10:44:13.891: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "dns-6825" for this suite. 02/25/23 10:44:13.994 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSS ------------------------------ -[sig-cli] Kubectl client Kubectl describe - should check if kubectl describe prints relevant information for rc and pods [Conformance] - test/e2e/kubectl/kubectl.go:1274 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:34:33.333 -Dec 14 16:34:33.333: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 16:34:33.335 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:33.363 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:33.366 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 -[It] should check if kubectl describe prints relevant information for rc and pods [Conformance] - test/e2e/kubectl/kubectl.go:1274 -Dec 14 16:34:33.369: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-546 create -f -' -Dec 14 16:34:34.934: INFO: stderr: "" -Dec 14 16:34:34.934: INFO: stdout: "replicationcontroller/agnhost-primary created\n" -Dec 14 16:34:34.935: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-546 create -f -' -Dec 14 16:34:36.500: INFO: stderr: "" -Dec 14 16:34:36.500: INFO: stdout: "service/agnhost-primary created\n" -STEP: Waiting for Agnhost primary to start. 12/14/22 16:34:36.5 -Dec 14 16:34:37.509: INFO: Selector matched 1 pods for map[app:agnhost] -Dec 14 16:34:37.509: INFO: Found 1 / 1 -Dec 14 16:34:37.509: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 -Dec 14 16:34:37.516: INFO: Selector matched 1 pods for map[app:agnhost] -Dec 14 16:34:37.516: INFO: ForEach: Found 1 pods from the filter. Now looping through them. -Dec 14 16:34:37.516: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-546 describe pod agnhost-primary-5g488' -Dec 14 16:34:37.655: INFO: stderr: "" -Dec 14 16:34:37.655: INFO: stdout: "Name: agnhost-primary-5g488\nNamespace: kubectl-546\nPriority: 0\nService Account: default\nNode: kiedai4iakae-3/192.168.121.244\nStart Time: Wed, 14 Dec 2022 16:34:34 +0000\nLabels: app=agnhost\n role=primary\nAnnotations: \nStatus: Running\nIP: 10.233.66.13\nIPs:\n IP: 10.233.66.13\nControlled By: ReplicationController/agnhost-primary\nContainers:\n agnhost-primary:\n Container ID: cri-o://1bc10193cc3cd58f2e8e079365ecfde5b397d6a6face038c1ce2912694a9f02c\n Image: registry.k8s.io/e2e-test-images/agnhost:2.40\n Image ID: registry.k8s.io/e2e-test-images/agnhost@sha256:37dd9ad84b34913b4bc57e00b0a28e8b00f18ca6a5ecec69461aa57402e5c787\n Port: 6379/TCP\n Host Port: 0/TCP\n State: Running\n Started: Wed, 14 Dec 2022 16:34:35 +0000\n Ready: True\n Restart Count: 0\n Environment: \n Mounts:\n /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-jgk2v (ro)\nConditions:\n Type Status\n Initialized True \n Ready True \n ContainersReady True \n PodScheduled True \nVolumes:\n kube-api-access-jgk2v:\n Type: Projected (a volume that contains injected data from multiple sources)\n TokenExpirationSeconds: 3607\n ConfigMapName: kube-root-ca.crt\n ConfigMapOptional: \n DownwardAPI: true\nQoS Class: BestEffort\nNode-Selectors: \nTolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s\n node.kubernetes.io/unreachable:NoExecute op=Exists for 300s\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal Scheduled 3s default-scheduler Successfully assigned kubectl-546/agnhost-primary-5g488 to kiedai4iakae-3\n Normal Pulled 2s kubelet Container image \"registry.k8s.io/e2e-test-images/agnhost:2.40\" already present on machine\n Normal Created 2s kubelet Created container agnhost-primary\n Normal Started 2s kubelet Started container agnhost-primary\n" -Dec 14 16:34:37.656: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-546 describe rc agnhost-primary' -Dec 14 16:34:37.810: INFO: stderr: "" -Dec 14 16:34:37.810: INFO: stdout: "Name: agnhost-primary\nNamespace: kubectl-546\nSelector: app=agnhost,role=primary\nLabels: app=agnhost\n role=primary\nAnnotations: \nReplicas: 1 current / 1 desired\nPods Status: 1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n Labels: app=agnhost\n role=primary\n Containers:\n agnhost-primary:\n Image: registry.k8s.io/e2e-test-images/agnhost:2.40\n Port: 6379/TCP\n Host Port: 0/TCP\n Environment: \n Mounts: \n Volumes: \nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal SuccessfulCreate 3s replication-controller Created pod: agnhost-primary-5g488\n" -Dec 14 16:34:37.811: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-546 describe service agnhost-primary' -Dec 14 16:34:37.965: INFO: stderr: "" -Dec 14 16:34:37.966: INFO: stdout: "Name: agnhost-primary\nNamespace: kubectl-546\nLabels: app=agnhost\n role=primary\nAnnotations: \nSelector: app=agnhost,role=primary\nType: ClusterIP\nIP Family Policy: SingleStack\nIP Families: IPv4\nIP: 10.233.12.157\nIPs: 10.233.12.157\nPort: 6379/TCP\nTargetPort: agnhost-server/TCP\nEndpoints: 10.233.66.13:6379\nSession Affinity: None\nEvents: \n" -Dec 14 16:34:37.971: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-546 describe node kiedai4iakae-1' -Dec 14 16:34:38.241: INFO: stderr: "" -Dec 14 16:34:38.241: INFO: stdout: "Name: kiedai4iakae-1\nRoles: control-plane\nLabels: beta.kubernetes.io/arch=amd64\n beta.kubernetes.io/os=linux\n kubernetes.io/arch=amd64\n kubernetes.io/hostname=kiedai4iakae-1\n kubernetes.io/os=linux\n node-role.kubernetes.io/control-plane=\n node.kubernetes.io/exclude-from-external-load-balancers=\nAnnotations: flannel.alpha.coreos.com/backend-data: {\"VNI\":1,\"VtepMAC\":\"8a:b0:d6:77:42:41\"}\n flannel.alpha.coreos.com/backend-type: vxlan\n flannel.alpha.coreos.com/kube-subnet-manager: true\n flannel.alpha.coreos.com/public-ip: 192.168.121.249\n kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/crio/crio.sock\n node.alpha.kubernetes.io/ttl: 0\n volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp: Wed, 14 Dec 2022 15:25:54 +0000\nTaints: \nUnschedulable: false\nLease:\n HolderIdentity: kiedai4iakae-1\n AcquireTime: \n RenewTime: Wed, 14 Dec 2022 16:34:29 +0000\nConditions:\n Type Status LastHeartbeatTime LastTransitionTime Reason Message\n ---- ------ ----------------- ------------------ ------ -------\n NetworkUnavailable False Wed, 14 Dec 2022 15:35:28 +0000 Wed, 14 Dec 2022 15:35:28 +0000 FlannelIsUp Flannel is running on this node\n MemoryPressure False Wed, 14 Dec 2022 16:31:11 +0000 Wed, 14 Dec 2022 15:25:46 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available\n DiskPressure False Wed, 14 Dec 2022 16:31:11 +0000 Wed, 14 Dec 2022 15:25:46 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure\n PIDPressure False Wed, 14 Dec 2022 16:31:11 +0000 Wed, 14 Dec 2022 15:25:46 +0000 KubeletHasSufficientPID kubelet has sufficient PID available\n Ready True Wed, 14 Dec 2022 16:31:11 +0000 Wed, 14 Dec 2022 15:29:20 +0000 KubeletReady kubelet is posting ready status. AppArmor enabled\nAddresses:\n InternalIP: 192.168.121.249\n Hostname: kiedai4iakae-1\nCapacity:\n cpu: 2\n ephemeral-storage: 122749536Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 8140760Ki\n pods: 110\n scheduling.k8s.io/foo: 5\nAllocatable:\n cpu: 1600m\n ephemeral-storage: 119410748528\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 3291096Ki\n pods: 110\n scheduling.k8s.io/foo: 5\nSystem Info:\n Machine ID: 079e0a7c4fda4d72b171fd8d77f59fd3\n System UUID: 079e0a7c-4fda-4d72-b171-fd8d77f59fd3\n Boot ID: c956f4e1-aebc-44b7-80de-9f3c20ca693e\n Kernel Version: 5.15.0-56-generic\n OS Image: Ubuntu 22.04.1 LTS\n Operating System: linux\n Architecture: amd64\n Container Runtime Version: cri-o://1.25.1\n Kubelet Version: v1.25.5\n Kube-Proxy Version: v1.25.5\nPodCIDR: 10.233.64.0/24\nPodCIDRs: 10.233.64.0/24\nNon-terminated Pods: (8 in total)\n Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age\n --------- ---- ------------ ---------- --------------- ------------- ---\n kube-flannel kube-flannel-ds-d6vqv 100m (6%) 100m (6%) 50Mi (1%) 50Mi (1%) 59m\n kube-system coredns-565d847f94-6mrkc 100m (6%) 0 (0%) 70Mi (2%) 170Mi (5%) 25m\n kube-system kube-addon-manager-kiedai4iakae-1 5m (0%) 0 (0%) 50Mi (1%) 0 (0%) 59m\n kube-system kube-apiserver-kiedai4iakae-1 250m (15%) 0 (0%) 0 (0%) 0 (0%) 59m\n kube-system kube-controller-manager-kiedai4iakae-1 200m (12%) 0 (0%) 0 (0%) 0 (0%) 59m\n kube-system kube-proxy-fs5fm 0 (0%) 0 (0%) 0 (0%) 0 (0%) 59m\n kube-system kube-scheduler-kiedai4iakae-1 100m (6%) 0 (0%) 0 (0%) 0 (0%) 59m\n sonobuoy sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-z7zks 0 (0%) 0 (0%) 0 (0%) 0 (0%) 58m\nAllocated resources:\n (Total limits may be over 100 percent, i.e., overcommitted.)\n Resource Requests Limits\n -------- -------- ------\n cpu 755m (47%) 100m (6%)\n memory 170Mi (5%) 220Mi (6%)\n ephemeral-storage 0 (0%) 0 (0%)\n hugepages-1Gi 0 (0%) 0 (0%)\n hugepages-2Mi 0 (0%) 0 (0%)\n scheduling.k8s.io/foo 0 0\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal Starting 68m kube-proxy \n Normal Starting 59m kube-proxy \n Normal Starting 59m kube-proxy \n Normal NodeHasSufficientMemory 68m (x6 over 68m) kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientMemory\n Normal NodeHasNoDiskPressure 68m (x6 over 68m) kubelet Node kiedai4iakae-1 status is now: NodeHasNoDiskPressure\n Normal NodeHasSufficientPID 68m (x6 over 68m) kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientPID\n Normal Starting 68m kubelet Starting kubelet.\n Normal NodeAllocatableEnforced 68m kubelet Updated Node Allocatable limit across pods\n Normal NodeHasNoDiskPressure 68m kubelet Node kiedai4iakae-1 status is now: NodeHasNoDiskPressure\n Normal NodeHasSufficientPID 68m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientPID\n Normal NodeHasSufficientMemory 68m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientMemory\n Normal NodeReady 68m kubelet Node kiedai4iakae-1 status is now: NodeReady\n Normal RegisteredNode 68m node-controller Node kiedai4iakae-1 event: Registered Node kiedai4iakae-1 in Controller\n Normal Starting 67m kubelet Starting kubelet.\n Normal NodeHasSufficientMemory 67m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientMemory\n Normal NodeHasNoDiskPressure 67m kubelet Node kiedai4iakae-1 status is now: NodeHasNoDiskPressure\n Normal NodeHasSufficientPID 67m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientPID\n Normal NodeNotReady 67m kubelet Node kiedai4iakae-1 status is now: NodeNotReady\n Normal NodeAllocatableEnforced 67m kubelet Updated Node Allocatable limit across pods\n Normal NodeReady 67m kubelet Node kiedai4iakae-1 status is now: NodeReady\n Normal Starting 65m kubelet Starting kubelet.\n Normal NodeHasSufficientMemory 65m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientMemory\n Normal NodeHasNoDiskPressure 65m kubelet Node kiedai4iakae-1 status is now: NodeHasNoDiskPressure\n Normal NodeHasSufficientPID 65m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientPID\n Normal NodeNotReady 65m kubelet Node kiedai4iakae-1 status is now: NodeNotReady\n Normal NodeAllocatableEnforced 65m kubelet Updated Node Allocatable limit across pods\n Normal NodeReady 65m kubelet Node kiedai4iakae-1 status is now: NodeReady\n Normal NodeNotReady 65m kubelet Node kiedai4iakae-1 status is now: NodeNotReady\n Normal NodeHasNoDiskPressure 65m kubelet Node kiedai4iakae-1 status is now: NodeHasNoDiskPressure\n Normal NodeHasSufficientPID 65m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientPID\n Normal Starting 65m kubelet Starting kubelet.\n Normal NodeAllocatableEnforced 65m kubelet Updated Node Allocatable limit across pods\n Normal NodeReady 65m kubelet Node kiedai4iakae-1 status is now: NodeReady\n Normal NodeHasSufficientMemory 65m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientMemory\n Normal RegisteredNode 61m node-controller Node kiedai4iakae-1 event: Registered Node kiedai4iakae-1 in Controller\n Normal Starting 59m kubelet Starting kubelet.\n Normal NodeAllocatableEnforced 59m kubelet Updated Node Allocatable limit across pods\n Normal NodeHasNoDiskPressure 59m (x8 over 59m) kubelet Node kiedai4iakae-1 status is now: NodeHasNoDiskPressure\n Normal NodeHasSufficientPID 59m (x7 over 59m) kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientPID\n Normal NodeHasSufficientMemory 59m (x8 over 59m) kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientMemory\n Normal RegisteredNode 59m node-controller Node kiedai4iakae-1 event: Registered Node kiedai4iakae-1 in Controller\n" -Dec 14 16:34:38.242: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-546 describe namespace kubectl-546' -Dec 14 16:34:38.403: INFO: stderr: "" -Dec 14 16:34:38.403: INFO: stdout: "Name: kubectl-546\nLabels: e2e-framework=kubectl\n e2e-run=211b00e9-8cf9-4510-82bc-c1b7016a99c3\n kubernetes.io/metadata.name=kubectl-546\n pod-security.kubernetes.io/enforce=baseline\nAnnotations: \nStatus: Active\n\nNo resource quota.\n\nNo LimitRange resource.\n" -[AfterEach] [sig-cli] Kubectl client +[sig-network] EndpointSlice + should have Endpoints and EndpointSlices pointing to API Server [Conformance] + test/e2e/network/endpointslice.go:65 +[BeforeEach] [sig-network] EndpointSlice + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:44:14.07 +Feb 25 10:44:14.070: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename endpointslice 02/25/23 10:44:14.079 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:44:14.131 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:44:14.141 +[BeforeEach] [sig-network] EndpointSlice + test/e2e/network/endpointslice.go:51 +[It] should have Endpoints and EndpointSlices pointing to API Server [Conformance] + test/e2e/network/endpointslice.go:65 +Feb 25 10:44:14.183: INFO: Endpoints addresses: [192.168.121.117 192.168.121.250] , ports: [6443] +Feb 25 10:44:14.183: INFO: EndpointSlices addresses: [192.168.121.117 192.168.121.250] , ports: [6443] +[AfterEach] [sig-network] EndpointSlice test/e2e/framework/framework.go:187 -Dec 14 16:34:38.403: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-546" for this suite. 12/14/22 16:34:38.419 -{"msg":"PASSED [sig-cli] Kubectl client Kubectl describe should check if kubectl describe prints relevant information for rc and pods [Conformance]","completed":181,"skipped":3348,"failed":0} +Feb 25 10:44:14.184: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "endpointslice-3849" for this suite. 02/25/23 10:44:14.193 +{"msg":"PASSED [sig-network] EndpointSlice should have Endpoints and EndpointSlices pointing to API Server [Conformance]","completed":177,"skipped":3396,"failed":0} ------------------------------ -• [SLOW TEST] [5.097 seconds] -[sig-cli] Kubectl client -test/e2e/kubectl/framework.go:23 - Kubectl describe - test/e2e/kubectl/kubectl.go:1268 - should check if kubectl describe prints relevant information for rc and pods [Conformance] - test/e2e/kubectl/kubectl.go:1274 +• [0.146 seconds] +[sig-network] EndpointSlice +test/e2e/network/common/framework.go:23 + should have Endpoints and EndpointSlices pointing to API Server [Conformance] + test/e2e/network/endpointslice.go:65 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-cli] Kubectl client + [BeforeEach] [sig-network] EndpointSlice test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:34:33.333 - Dec 14 16:34:33.333: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 16:34:33.335 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:33.363 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:33.366 - [BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 - [It] should check if kubectl describe prints relevant information for rc and pods [Conformance] - test/e2e/kubectl/kubectl.go:1274 - Dec 14 16:34:33.369: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-546 create -f -' - Dec 14 16:34:34.934: INFO: stderr: "" - Dec 14 16:34:34.934: INFO: stdout: "replicationcontroller/agnhost-primary created\n" - Dec 14 16:34:34.935: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-546 create -f -' - Dec 14 16:34:36.500: INFO: stderr: "" - Dec 14 16:34:36.500: INFO: stdout: "service/agnhost-primary created\n" - STEP: Waiting for Agnhost primary to start. 12/14/22 16:34:36.5 - Dec 14 16:34:37.509: INFO: Selector matched 1 pods for map[app:agnhost] - Dec 14 16:34:37.509: INFO: Found 1 / 1 - Dec 14 16:34:37.509: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 - Dec 14 16:34:37.516: INFO: Selector matched 1 pods for map[app:agnhost] - Dec 14 16:34:37.516: INFO: ForEach: Found 1 pods from the filter. Now looping through them. - Dec 14 16:34:37.516: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-546 describe pod agnhost-primary-5g488' - Dec 14 16:34:37.655: INFO: stderr: "" - Dec 14 16:34:37.655: INFO: stdout: "Name: agnhost-primary-5g488\nNamespace: kubectl-546\nPriority: 0\nService Account: default\nNode: kiedai4iakae-3/192.168.121.244\nStart Time: Wed, 14 Dec 2022 16:34:34 +0000\nLabels: app=agnhost\n role=primary\nAnnotations: \nStatus: Running\nIP: 10.233.66.13\nIPs:\n IP: 10.233.66.13\nControlled By: ReplicationController/agnhost-primary\nContainers:\n agnhost-primary:\n Container ID: cri-o://1bc10193cc3cd58f2e8e079365ecfde5b397d6a6face038c1ce2912694a9f02c\n Image: registry.k8s.io/e2e-test-images/agnhost:2.40\n Image ID: registry.k8s.io/e2e-test-images/agnhost@sha256:37dd9ad84b34913b4bc57e00b0a28e8b00f18ca6a5ecec69461aa57402e5c787\n Port: 6379/TCP\n Host Port: 0/TCP\n State: Running\n Started: Wed, 14 Dec 2022 16:34:35 +0000\n Ready: True\n Restart Count: 0\n Environment: \n Mounts:\n /var/run/secrets/kubernetes.io/serviceaccount from kube-api-access-jgk2v (ro)\nConditions:\n Type Status\n Initialized True \n Ready True \n ContainersReady True \n PodScheduled True \nVolumes:\n kube-api-access-jgk2v:\n Type: Projected (a volume that contains injected data from multiple sources)\n TokenExpirationSeconds: 3607\n ConfigMapName: kube-root-ca.crt\n ConfigMapOptional: \n DownwardAPI: true\nQoS Class: BestEffort\nNode-Selectors: \nTolerations: node.kubernetes.io/not-ready:NoExecute op=Exists for 300s\n node.kubernetes.io/unreachable:NoExecute op=Exists for 300s\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal Scheduled 3s default-scheduler Successfully assigned kubectl-546/agnhost-primary-5g488 to kiedai4iakae-3\n Normal Pulled 2s kubelet Container image \"registry.k8s.io/e2e-test-images/agnhost:2.40\" already present on machine\n Normal Created 2s kubelet Created container agnhost-primary\n Normal Started 2s kubelet Started container agnhost-primary\n" - Dec 14 16:34:37.656: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-546 describe rc agnhost-primary' - Dec 14 16:34:37.810: INFO: stderr: "" - Dec 14 16:34:37.810: INFO: stdout: "Name: agnhost-primary\nNamespace: kubectl-546\nSelector: app=agnhost,role=primary\nLabels: app=agnhost\n role=primary\nAnnotations: \nReplicas: 1 current / 1 desired\nPods Status: 1 Running / 0 Waiting / 0 Succeeded / 0 Failed\nPod Template:\n Labels: app=agnhost\n role=primary\n Containers:\n agnhost-primary:\n Image: registry.k8s.io/e2e-test-images/agnhost:2.40\n Port: 6379/TCP\n Host Port: 0/TCP\n Environment: \n Mounts: \n Volumes: \nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal SuccessfulCreate 3s replication-controller Created pod: agnhost-primary-5g488\n" - Dec 14 16:34:37.811: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-546 describe service agnhost-primary' - Dec 14 16:34:37.965: INFO: stderr: "" - Dec 14 16:34:37.966: INFO: stdout: "Name: agnhost-primary\nNamespace: kubectl-546\nLabels: app=agnhost\n role=primary\nAnnotations: \nSelector: app=agnhost,role=primary\nType: ClusterIP\nIP Family Policy: SingleStack\nIP Families: IPv4\nIP: 10.233.12.157\nIPs: 10.233.12.157\nPort: 6379/TCP\nTargetPort: agnhost-server/TCP\nEndpoints: 10.233.66.13:6379\nSession Affinity: None\nEvents: \n" - Dec 14 16:34:37.971: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-546 describe node kiedai4iakae-1' - Dec 14 16:34:38.241: INFO: stderr: "" - Dec 14 16:34:38.241: INFO: stdout: "Name: kiedai4iakae-1\nRoles: control-plane\nLabels: beta.kubernetes.io/arch=amd64\n beta.kubernetes.io/os=linux\n kubernetes.io/arch=amd64\n kubernetes.io/hostname=kiedai4iakae-1\n kubernetes.io/os=linux\n node-role.kubernetes.io/control-plane=\n node.kubernetes.io/exclude-from-external-load-balancers=\nAnnotations: flannel.alpha.coreos.com/backend-data: {\"VNI\":1,\"VtepMAC\":\"8a:b0:d6:77:42:41\"}\n flannel.alpha.coreos.com/backend-type: vxlan\n flannel.alpha.coreos.com/kube-subnet-manager: true\n flannel.alpha.coreos.com/public-ip: 192.168.121.249\n kubeadm.alpha.kubernetes.io/cri-socket: unix:///var/run/crio/crio.sock\n node.alpha.kubernetes.io/ttl: 0\n volumes.kubernetes.io/controller-managed-attach-detach: true\nCreationTimestamp: Wed, 14 Dec 2022 15:25:54 +0000\nTaints: \nUnschedulable: false\nLease:\n HolderIdentity: kiedai4iakae-1\n AcquireTime: \n RenewTime: Wed, 14 Dec 2022 16:34:29 +0000\nConditions:\n Type Status LastHeartbeatTime LastTransitionTime Reason Message\n ---- ------ ----------------- ------------------ ------ -------\n NetworkUnavailable False Wed, 14 Dec 2022 15:35:28 +0000 Wed, 14 Dec 2022 15:35:28 +0000 FlannelIsUp Flannel is running on this node\n MemoryPressure False Wed, 14 Dec 2022 16:31:11 +0000 Wed, 14 Dec 2022 15:25:46 +0000 KubeletHasSufficientMemory kubelet has sufficient memory available\n DiskPressure False Wed, 14 Dec 2022 16:31:11 +0000 Wed, 14 Dec 2022 15:25:46 +0000 KubeletHasNoDiskPressure kubelet has no disk pressure\n PIDPressure False Wed, 14 Dec 2022 16:31:11 +0000 Wed, 14 Dec 2022 15:25:46 +0000 KubeletHasSufficientPID kubelet has sufficient PID available\n Ready True Wed, 14 Dec 2022 16:31:11 +0000 Wed, 14 Dec 2022 15:29:20 +0000 KubeletReady kubelet is posting ready status. AppArmor enabled\nAddresses:\n InternalIP: 192.168.121.249\n Hostname: kiedai4iakae-1\nCapacity:\n cpu: 2\n ephemeral-storage: 122749536Ki\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 8140760Ki\n pods: 110\n scheduling.k8s.io/foo: 5\nAllocatable:\n cpu: 1600m\n ephemeral-storage: 119410748528\n hugepages-1Gi: 0\n hugepages-2Mi: 0\n memory: 3291096Ki\n pods: 110\n scheduling.k8s.io/foo: 5\nSystem Info:\n Machine ID: 079e0a7c4fda4d72b171fd8d77f59fd3\n System UUID: 079e0a7c-4fda-4d72-b171-fd8d77f59fd3\n Boot ID: c956f4e1-aebc-44b7-80de-9f3c20ca693e\n Kernel Version: 5.15.0-56-generic\n OS Image: Ubuntu 22.04.1 LTS\n Operating System: linux\n Architecture: amd64\n Container Runtime Version: cri-o://1.25.1\n Kubelet Version: v1.25.5\n Kube-Proxy Version: v1.25.5\nPodCIDR: 10.233.64.0/24\nPodCIDRs: 10.233.64.0/24\nNon-terminated Pods: (8 in total)\n Namespace Name CPU Requests CPU Limits Memory Requests Memory Limits Age\n --------- ---- ------------ ---------- --------------- ------------- ---\n kube-flannel kube-flannel-ds-d6vqv 100m (6%) 100m (6%) 50Mi (1%) 50Mi (1%) 59m\n kube-system coredns-565d847f94-6mrkc 100m (6%) 0 (0%) 70Mi (2%) 170Mi (5%) 25m\n kube-system kube-addon-manager-kiedai4iakae-1 5m (0%) 0 (0%) 50Mi (1%) 0 (0%) 59m\n kube-system kube-apiserver-kiedai4iakae-1 250m (15%) 0 (0%) 0 (0%) 0 (0%) 59m\n kube-system kube-controller-manager-kiedai4iakae-1 200m (12%) 0 (0%) 0 (0%) 0 (0%) 59m\n kube-system kube-proxy-fs5fm 0 (0%) 0 (0%) 0 (0%) 0 (0%) 59m\n kube-system kube-scheduler-kiedai4iakae-1 100m (6%) 0 (0%) 0 (0%) 0 (0%) 59m\n sonobuoy sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-z7zks 0 (0%) 0 (0%) 0 (0%) 0 (0%) 58m\nAllocated resources:\n (Total limits may be over 100 percent, i.e., overcommitted.)\n Resource Requests Limits\n -------- -------- ------\n cpu 755m (47%) 100m (6%)\n memory 170Mi (5%) 220Mi (6%)\n ephemeral-storage 0 (0%) 0 (0%)\n hugepages-1Gi 0 (0%) 0 (0%)\n hugepages-2Mi 0 (0%) 0 (0%)\n scheduling.k8s.io/foo 0 0\nEvents:\n Type Reason Age From Message\n ---- ------ ---- ---- -------\n Normal Starting 68m kube-proxy \n Normal Starting 59m kube-proxy \n Normal Starting 59m kube-proxy \n Normal NodeHasSufficientMemory 68m (x6 over 68m) kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientMemory\n Normal NodeHasNoDiskPressure 68m (x6 over 68m) kubelet Node kiedai4iakae-1 status is now: NodeHasNoDiskPressure\n Normal NodeHasSufficientPID 68m (x6 over 68m) kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientPID\n Normal Starting 68m kubelet Starting kubelet.\n Normal NodeAllocatableEnforced 68m kubelet Updated Node Allocatable limit across pods\n Normal NodeHasNoDiskPressure 68m kubelet Node kiedai4iakae-1 status is now: NodeHasNoDiskPressure\n Normal NodeHasSufficientPID 68m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientPID\n Normal NodeHasSufficientMemory 68m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientMemory\n Normal NodeReady 68m kubelet Node kiedai4iakae-1 status is now: NodeReady\n Normal RegisteredNode 68m node-controller Node kiedai4iakae-1 event: Registered Node kiedai4iakae-1 in Controller\n Normal Starting 67m kubelet Starting kubelet.\n Normal NodeHasSufficientMemory 67m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientMemory\n Normal NodeHasNoDiskPressure 67m kubelet Node kiedai4iakae-1 status is now: NodeHasNoDiskPressure\n Normal NodeHasSufficientPID 67m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientPID\n Normal NodeNotReady 67m kubelet Node kiedai4iakae-1 status is now: NodeNotReady\n Normal NodeAllocatableEnforced 67m kubelet Updated Node Allocatable limit across pods\n Normal NodeReady 67m kubelet Node kiedai4iakae-1 status is now: NodeReady\n Normal Starting 65m kubelet Starting kubelet.\n Normal NodeHasSufficientMemory 65m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientMemory\n Normal NodeHasNoDiskPressure 65m kubelet Node kiedai4iakae-1 status is now: NodeHasNoDiskPressure\n Normal NodeHasSufficientPID 65m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientPID\n Normal NodeNotReady 65m kubelet Node kiedai4iakae-1 status is now: NodeNotReady\n Normal NodeAllocatableEnforced 65m kubelet Updated Node Allocatable limit across pods\n Normal NodeReady 65m kubelet Node kiedai4iakae-1 status is now: NodeReady\n Normal NodeNotReady 65m kubelet Node kiedai4iakae-1 status is now: NodeNotReady\n Normal NodeHasNoDiskPressure 65m kubelet Node kiedai4iakae-1 status is now: NodeHasNoDiskPressure\n Normal NodeHasSufficientPID 65m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientPID\n Normal Starting 65m kubelet Starting kubelet.\n Normal NodeAllocatableEnforced 65m kubelet Updated Node Allocatable limit across pods\n Normal NodeReady 65m kubelet Node kiedai4iakae-1 status is now: NodeReady\n Normal NodeHasSufficientMemory 65m kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientMemory\n Normal RegisteredNode 61m node-controller Node kiedai4iakae-1 event: Registered Node kiedai4iakae-1 in Controller\n Normal Starting 59m kubelet Starting kubelet.\n Normal NodeAllocatableEnforced 59m kubelet Updated Node Allocatable limit across pods\n Normal NodeHasNoDiskPressure 59m (x8 over 59m) kubelet Node kiedai4iakae-1 status is now: NodeHasNoDiskPressure\n Normal NodeHasSufficientPID 59m (x7 over 59m) kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientPID\n Normal NodeHasSufficientMemory 59m (x8 over 59m) kubelet Node kiedai4iakae-1 status is now: NodeHasSufficientMemory\n Normal RegisteredNode 59m node-controller Node kiedai4iakae-1 event: Registered Node kiedai4iakae-1 in Controller\n" - Dec 14 16:34:38.242: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-546 describe namespace kubectl-546' - Dec 14 16:34:38.403: INFO: stderr: "" - Dec 14 16:34:38.403: INFO: stdout: "Name: kubectl-546\nLabels: e2e-framework=kubectl\n e2e-run=211b00e9-8cf9-4510-82bc-c1b7016a99c3\n kubernetes.io/metadata.name=kubectl-546\n pod-security.kubernetes.io/enforce=baseline\nAnnotations: \nStatus: Active\n\nNo resource quota.\n\nNo LimitRange resource.\n" - [AfterEach] [sig-cli] Kubectl client + STEP: Creating a kubernetes client 02/25/23 10:44:14.07 + Feb 25 10:44:14.070: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename endpointslice 02/25/23 10:44:14.079 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:44:14.131 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:44:14.141 + [BeforeEach] [sig-network] EndpointSlice + test/e2e/network/endpointslice.go:51 + [It] should have Endpoints and EndpointSlices pointing to API Server [Conformance] + test/e2e/network/endpointslice.go:65 + Feb 25 10:44:14.183: INFO: Endpoints addresses: [192.168.121.117 192.168.121.250] , ports: [6443] + Feb 25 10:44:14.183: INFO: EndpointSlices addresses: [192.168.121.117 192.168.121.250] , ports: [6443] + [AfterEach] [sig-network] EndpointSlice test/e2e/framework/framework.go:187 - Dec 14 16:34:38.403: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-546" for this suite. 12/14/22 16:34:38.419 + Feb 25 10:44:14.184: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "endpointslice-3849" for this suite. 02/25/23 10:44:14.193 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSS ------------------------------ -[sig-storage] Projected secret - should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:55 -[BeforeEach] [sig-storage] Projected secret +[sig-storage] ConfigMap + optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:239 +[BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:34:38.438 -Dec 14 16:34:38.438: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 16:34:38.442 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:38.476 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:38.48 -[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:55 -STEP: Creating projection with secret that has name projected-secret-test-71849257-e739-4043-b994-a60ee081e3af 12/14/22 16:34:38.484 -STEP: Creating a pod to test consume secrets 12/14/22 16:34:38.496 -Dec 14 16:34:38.522: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661" in namespace "projected-765" to be "Succeeded or Failed" -Dec 14 16:34:38.527: INFO: Pod "pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661": Phase="Pending", Reason="", readiness=false. Elapsed: 4.846495ms -Dec 14 16:34:40.533: INFO: Pod "pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011320759s -Dec 14 16:34:42.533: INFO: Pod "pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010804346s -STEP: Saw pod success 12/14/22 16:34:42.533 -Dec 14 16:34:42.533: INFO: Pod "pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661" satisfied condition "Succeeded or Failed" -Dec 14 16:34:42.538: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661 container projected-secret-volume-test: -STEP: delete the pod 12/14/22 16:34:42.551 -Dec 14 16:34:42.584: INFO: Waiting for pod pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661 to disappear -Dec 14 16:34:42.589: INFO: Pod pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661 no longer exists -[AfterEach] [sig-storage] Projected secret +STEP: Creating a kubernetes client 02/25/23 10:44:14.221 +Feb 25 10:44:14.221: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename configmap 02/25/23 10:44:14.225 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:44:14.261 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:44:14.268 +[It] optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:239 +STEP: Creating configMap with name cm-test-opt-del-e2c1fd7d-320b-4a83-9d35-9f909a0b3cbe 02/25/23 10:44:14.29 +STEP: Creating configMap with name cm-test-opt-upd-d3ab535d-9818-4787-aa41-5826d2d27d00 02/25/23 10:44:14.32 +STEP: Creating the pod 02/25/23 10:44:14.332 +Feb 25 10:44:14.357: INFO: Waiting up to 5m0s for pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562" in namespace "configmap-6434" to be "running and ready" +Feb 25 10:44:14.367: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 9.439535ms +Feb 25 10:44:14.367: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:44:16.376: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018596098s +Feb 25 10:44:16.376: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:44:18.377: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 4.019720863s +Feb 25 10:44:18.377: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:44:20.380: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 6.023111285s +Feb 25 10:44:20.380: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:44:22.376: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 8.018385579s +Feb 25 10:44:22.376: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:44:24.379: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 10.021616014s +Feb 25 10:44:24.379: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:44:26.377: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 12.019413516s +Feb 25 10:44:26.377: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:44:28.436: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 14.078838489s +Feb 25 10:44:28.444: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:44:30.381: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 16.023371543s +Feb 25 10:44:30.381: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:44:32.379: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Running", Reason="", readiness=true. Elapsed: 18.02182254s +Feb 25 10:44:32.379: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Running (Ready = true) +Feb 25 10:44:32.379: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562" satisfied condition "running and ready" +STEP: Deleting configmap cm-test-opt-del-e2c1fd7d-320b-4a83-9d35-9f909a0b3cbe 02/25/23 10:44:32.459 +STEP: Updating configmap cm-test-opt-upd-d3ab535d-9818-4787-aa41-5826d2d27d00 02/25/23 10:44:32.478 +STEP: Creating configMap with name cm-test-opt-create-c740cd6b-aa01-4295-b51d-09c5e09aac66 02/25/23 10:44:32.489 +STEP: waiting to observe update in volume 02/25/23 10:44:32.499 +[AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 -Dec 14 16:34:42.590: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-765" for this suite. 12/14/22 16:34:42.597 -{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","completed":182,"skipped":3373,"failed":0} +Feb 25 10:46:04.372: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-6434" for this suite. 02/25/23 10:46:04.385 +{"msg":"PASSED [sig-storage] ConfigMap optional updates should be reflected in volume [NodeConformance] [Conformance]","completed":178,"skipped":3404,"failed":0} ------------------------------ -• [4.173 seconds] -[sig-storage] Projected secret +• [SLOW TEST] [110.180 seconds] +[sig-storage] ConfigMap test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:55 + optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:239 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected secret + [BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:34:38.438 - Dec 14 16:34:38.438: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 16:34:38.442 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:38.476 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:38.48 - [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:55 - STEP: Creating projection with secret that has name projected-secret-test-71849257-e739-4043-b994-a60ee081e3af 12/14/22 16:34:38.484 - STEP: Creating a pod to test consume secrets 12/14/22 16:34:38.496 - Dec 14 16:34:38.522: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661" in namespace "projected-765" to be "Succeeded or Failed" - Dec 14 16:34:38.527: INFO: Pod "pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661": Phase="Pending", Reason="", readiness=false. Elapsed: 4.846495ms - Dec 14 16:34:40.533: INFO: Pod "pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011320759s - Dec 14 16:34:42.533: INFO: Pod "pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010804346s - STEP: Saw pod success 12/14/22 16:34:42.533 - Dec 14 16:34:42.533: INFO: Pod "pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661" satisfied condition "Succeeded or Failed" - Dec 14 16:34:42.538: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661 container projected-secret-volume-test: - STEP: delete the pod 12/14/22 16:34:42.551 - Dec 14 16:34:42.584: INFO: Waiting for pod pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661 to disappear - Dec 14 16:34:42.589: INFO: Pod pod-projected-secrets-ba41270b-9e5e-4ae7-acc7-9345d95ba661 no longer exists - [AfterEach] [sig-storage] Projected secret + STEP: Creating a kubernetes client 02/25/23 10:44:14.221 + Feb 25 10:44:14.221: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename configmap 02/25/23 10:44:14.225 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:44:14.261 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:44:14.268 + [It] optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:239 + STEP: Creating configMap with name cm-test-opt-del-e2c1fd7d-320b-4a83-9d35-9f909a0b3cbe 02/25/23 10:44:14.29 + STEP: Creating configMap with name cm-test-opt-upd-d3ab535d-9818-4787-aa41-5826d2d27d00 02/25/23 10:44:14.32 + STEP: Creating the pod 02/25/23 10:44:14.332 + Feb 25 10:44:14.357: INFO: Waiting up to 5m0s for pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562" in namespace "configmap-6434" to be "running and ready" + Feb 25 10:44:14.367: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 9.439535ms + Feb 25 10:44:14.367: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:44:16.376: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018596098s + Feb 25 10:44:16.376: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:44:18.377: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 4.019720863s + Feb 25 10:44:18.377: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:44:20.380: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 6.023111285s + Feb 25 10:44:20.380: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:44:22.376: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 8.018385579s + Feb 25 10:44:22.376: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:44:24.379: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 10.021616014s + Feb 25 10:44:24.379: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:44:26.377: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 12.019413516s + Feb 25 10:44:26.377: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:44:28.436: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 14.078838489s + Feb 25 10:44:28.444: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:44:30.381: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Pending", Reason="", readiness=false. Elapsed: 16.023371543s + Feb 25 10:44:30.381: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:44:32.379: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562": Phase="Running", Reason="", readiness=true. Elapsed: 18.02182254s + Feb 25 10:44:32.379: INFO: The phase of Pod pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562 is Running (Ready = true) + Feb 25 10:44:32.379: INFO: Pod "pod-configmaps-82de2e17-7675-43f9-8b62-ccbf804b4562" satisfied condition "running and ready" + STEP: Deleting configmap cm-test-opt-del-e2c1fd7d-320b-4a83-9d35-9f909a0b3cbe 02/25/23 10:44:32.459 + STEP: Updating configmap cm-test-opt-upd-d3ab535d-9818-4787-aa41-5826d2d27d00 02/25/23 10:44:32.478 + STEP: Creating configMap with name cm-test-opt-create-c740cd6b-aa01-4295-b51d-09c5e09aac66 02/25/23 10:44:32.489 + STEP: waiting to observe update in volume 02/25/23 10:44:32.499 + [AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 - Dec 14 16:34:42.590: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-765" for this suite. 12/14/22 16:34:42.597 + Feb 25 10:46:04.372: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "configmap-6434" for this suite. 02/25/23 10:46:04.385 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSSSSSS ------------------------------ -[sig-network] Services - should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance] - test/e2e/network/service.go:2237 -[BeforeEach] [sig-network] Services +[sig-apps] CronJob + should support CronJob API operations [Conformance] + test/e2e/apps/cronjob.go:319 +[BeforeEach] [sig-apps] CronJob test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:34:42.614 -Dec 14 16:34:42.614: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 16:34:42.616 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:42.639 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:42.645 -[BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 -[It] should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance] - test/e2e/network/service.go:2237 -STEP: creating service in namespace services-6337 12/14/22 16:34:42.649 -STEP: creating service affinity-nodeport-transition in namespace services-6337 12/14/22 16:34:42.65 -STEP: creating replication controller affinity-nodeport-transition in namespace services-6337 12/14/22 16:34:42.677 -I1214 16:34:42.692428 14 runners.go:193] Created replication controller with name: affinity-nodeport-transition, namespace: services-6337, replica count: 3 -I1214 16:34:45.745139 14 runners.go:193] affinity-nodeport-transition Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -Dec 14 16:34:45.770: INFO: Creating new exec pod -Dec 14 16:34:45.790: INFO: Waiting up to 5m0s for pod "execpod-affinityr4wms" in namespace "services-6337" to be "running" -Dec 14 16:34:45.794: INFO: Pod "execpod-affinityr4wms": Phase="Pending", Reason="", readiness=false. Elapsed: 3.551734ms -Dec 14 16:34:47.800: INFO: Pod "execpod-affinityr4wms": Phase="Pending", Reason="", readiness=false. Elapsed: 2.009460472s -Dec 14 16:34:49.801: INFO: Pod "execpod-affinityr4wms": Phase="Running", Reason="", readiness=true. Elapsed: 4.011010676s -Dec 14 16:34:49.801: INFO: Pod "execpod-affinityr4wms" satisfied condition "running" -Dec 14 16:34:50.810: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-6337 exec execpod-affinityr4wms -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-nodeport-transition 80' -Dec 14 16:34:51.069: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-nodeport-transition 80\nConnection to affinity-nodeport-transition 80 port [tcp/http] succeeded!\n" -Dec 14 16:34:51.069: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:34:51.069: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-6337 exec execpod-affinityr4wms -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.36.193 80' -Dec 14 16:34:51.299: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.36.193 80\nConnection to 10.233.36.193 80 port [tcp/http] succeeded!\n" -Dec 14 16:34:51.299: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:34:51.300: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-6337 exec execpod-affinityr4wms -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.244 31588' -Dec 14 16:34:51.507: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.244 31588\nConnection to 192.168.121.244 31588 port [tcp/*] succeeded!\n" -Dec 14 16:34:51.507: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:34:51.508: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-6337 exec execpod-affinityr4wms -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.249 31588' -Dec 14 16:34:51.727: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.249 31588\nConnection to 192.168.121.249 31588 port [tcp/*] succeeded!\n" -Dec 14 16:34:51.727: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:34:51.746: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-6337 exec execpod-affinityr4wms -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.249:31588/ ; done' -Dec 14 16:34:52.515: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n" -Dec 14 16:34:52.516: INFO: stdout: "\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-k2fx9\naffinity-nodeport-transition-k2fx9\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-ck5sv\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-k2fx9\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-k2fx9\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-k2fx9\naffinity-nodeport-transition-ck5sv\naffinity-nodeport-transition-pst4p" -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-k2fx9 -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-k2fx9 -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-ck5sv -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-k2fx9 -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-k2fx9 -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-k2fx9 -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-ck5sv -Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:52.528: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-6337 exec execpod-affinityr4wms -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.249:31588/ ; done' -Dec 14 16:34:53.069: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n" -Dec 14 16:34:53.069: INFO: stdout: "\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p" -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p -Dec 14 16:34:53.069: INFO: Cleaning up the exec pod -STEP: deleting ReplicationController affinity-nodeport-transition in namespace services-6337, will wait for the garbage collector to delete the pods 12/14/22 16:34:53.121 -Dec 14 16:34:53.201: INFO: Deleting ReplicationController affinity-nodeport-transition took: 10.833382ms -Dec 14 16:34:53.302: INFO: Terminating ReplicationController affinity-nodeport-transition pods took: 100.689473ms -[AfterEach] [sig-network] Services +STEP: Creating a kubernetes client 02/25/23 10:46:04.415 +Feb 25 10:46:04.416: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename cronjob 02/25/23 10:46:04.424 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:04.464 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:04.474 +[It] should support CronJob API operations [Conformance] + test/e2e/apps/cronjob.go:319 +STEP: Creating a cronjob 02/25/23 10:46:04.485 +STEP: creating 02/25/23 10:46:04.486 +STEP: getting 02/25/23 10:46:04.506 +STEP: listing 02/25/23 10:46:04.514 +STEP: watching 02/25/23 10:46:04.521 +Feb 25 10:46:04.522: INFO: starting watch +STEP: cluster-wide listing 02/25/23 10:46:04.524 +STEP: cluster-wide watching 02/25/23 10:46:04.535 +Feb 25 10:46:04.536: INFO: starting watch +STEP: patching 02/25/23 10:46:04.538 +STEP: updating 02/25/23 10:46:04.551 +Feb 25 10:46:04.576: INFO: waiting for watch events with expected annotations +Feb 25 10:46:04.577: INFO: saw patched and updated annotations +STEP: patching /status 02/25/23 10:46:04.577 +STEP: updating /status 02/25/23 10:46:04.595 +STEP: get /status 02/25/23 10:46:04.623 +STEP: deleting 02/25/23 10:46:04.632 +STEP: deleting a collection 02/25/23 10:46:04.667 +[AfterEach] [sig-apps] CronJob test/e2e/framework/framework.go:187 -Dec 14 16:34:55.445: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-6337" for this suite. 12/14/22 16:34:55.451 -[AfterEach] [sig-network] Services - test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance]","completed":183,"skipped":3376,"failed":0} +Feb 25 10:46:04.709: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "cronjob-3574" for this suite. 02/25/23 10:46:04.72 +{"msg":"PASSED [sig-apps] CronJob should support CronJob API operations [Conformance]","completed":179,"skipped":3414,"failed":0} ------------------------------ -• [SLOW TEST] [12.850 seconds] -[sig-network] Services -test/e2e/network/common/framework.go:23 - should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance] - test/e2e/network/service.go:2237 +• [0.320 seconds] +[sig-apps] CronJob +test/e2e/apps/framework.go:23 + should support CronJob API operations [Conformance] + test/e2e/apps/cronjob.go:319 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Services + [BeforeEach] [sig-apps] CronJob test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:34:42.614 - Dec 14 16:34:42.614: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 16:34:42.616 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:42.639 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:42.645 - [BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 - [It] should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance] - test/e2e/network/service.go:2237 - STEP: creating service in namespace services-6337 12/14/22 16:34:42.649 - STEP: creating service affinity-nodeport-transition in namespace services-6337 12/14/22 16:34:42.65 - STEP: creating replication controller affinity-nodeport-transition in namespace services-6337 12/14/22 16:34:42.677 - I1214 16:34:42.692428 14 runners.go:193] Created replication controller with name: affinity-nodeport-transition, namespace: services-6337, replica count: 3 - I1214 16:34:45.745139 14 runners.go:193] affinity-nodeport-transition Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - Dec 14 16:34:45.770: INFO: Creating new exec pod - Dec 14 16:34:45.790: INFO: Waiting up to 5m0s for pod "execpod-affinityr4wms" in namespace "services-6337" to be "running" - Dec 14 16:34:45.794: INFO: Pod "execpod-affinityr4wms": Phase="Pending", Reason="", readiness=false. Elapsed: 3.551734ms - Dec 14 16:34:47.800: INFO: Pod "execpod-affinityr4wms": Phase="Pending", Reason="", readiness=false. Elapsed: 2.009460472s - Dec 14 16:34:49.801: INFO: Pod "execpod-affinityr4wms": Phase="Running", Reason="", readiness=true. Elapsed: 4.011010676s - Dec 14 16:34:49.801: INFO: Pod "execpod-affinityr4wms" satisfied condition "running" - Dec 14 16:34:50.810: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-6337 exec execpod-affinityr4wms -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-nodeport-transition 80' - Dec 14 16:34:51.069: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-nodeport-transition 80\nConnection to affinity-nodeport-transition 80 port [tcp/http] succeeded!\n" - Dec 14 16:34:51.069: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:34:51.069: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-6337 exec execpod-affinityr4wms -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.36.193 80' - Dec 14 16:34:51.299: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.36.193 80\nConnection to 10.233.36.193 80 port [tcp/http] succeeded!\n" - Dec 14 16:34:51.299: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:34:51.300: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-6337 exec execpod-affinityr4wms -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.244 31588' - Dec 14 16:34:51.507: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.244 31588\nConnection to 192.168.121.244 31588 port [tcp/*] succeeded!\n" - Dec 14 16:34:51.507: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:34:51.508: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-6337 exec execpod-affinityr4wms -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.249 31588' - Dec 14 16:34:51.727: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.249 31588\nConnection to 192.168.121.249 31588 port [tcp/*] succeeded!\n" - Dec 14 16:34:51.727: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:34:51.746: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-6337 exec execpod-affinityr4wms -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.249:31588/ ; done' - Dec 14 16:34:52.515: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n" - Dec 14 16:34:52.516: INFO: stdout: "\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-k2fx9\naffinity-nodeport-transition-k2fx9\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-ck5sv\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-k2fx9\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-k2fx9\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-k2fx9\naffinity-nodeport-transition-ck5sv\naffinity-nodeport-transition-pst4p" - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-k2fx9 - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-k2fx9 - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-ck5sv - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-k2fx9 - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-k2fx9 - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-k2fx9 - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-ck5sv - Dec 14 16:34:52.516: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:52.528: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-6337 exec execpod-affinityr4wms -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.249:31588/ ; done' - Dec 14 16:34:53.069: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:31588/\n" - Dec 14 16:34:53.069: INFO: stdout: "\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p\naffinity-nodeport-transition-pst4p" - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Received response from host: affinity-nodeport-transition-pst4p - Dec 14 16:34:53.069: INFO: Cleaning up the exec pod - STEP: deleting ReplicationController affinity-nodeport-transition in namespace services-6337, will wait for the garbage collector to delete the pods 12/14/22 16:34:53.121 - Dec 14 16:34:53.201: INFO: Deleting ReplicationController affinity-nodeport-transition took: 10.833382ms - Dec 14 16:34:53.302: INFO: Terminating ReplicationController affinity-nodeport-transition pods took: 100.689473ms - [AfterEach] [sig-network] Services + STEP: Creating a kubernetes client 02/25/23 10:46:04.415 + Feb 25 10:46:04.416: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename cronjob 02/25/23 10:46:04.424 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:04.464 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:04.474 + [It] should support CronJob API operations [Conformance] + test/e2e/apps/cronjob.go:319 + STEP: Creating a cronjob 02/25/23 10:46:04.485 + STEP: creating 02/25/23 10:46:04.486 + STEP: getting 02/25/23 10:46:04.506 + STEP: listing 02/25/23 10:46:04.514 + STEP: watching 02/25/23 10:46:04.521 + Feb 25 10:46:04.522: INFO: starting watch + STEP: cluster-wide listing 02/25/23 10:46:04.524 + STEP: cluster-wide watching 02/25/23 10:46:04.535 + Feb 25 10:46:04.536: INFO: starting watch + STEP: patching 02/25/23 10:46:04.538 + STEP: updating 02/25/23 10:46:04.551 + Feb 25 10:46:04.576: INFO: waiting for watch events with expected annotations + Feb 25 10:46:04.577: INFO: saw patched and updated annotations + STEP: patching /status 02/25/23 10:46:04.577 + STEP: updating /status 02/25/23 10:46:04.595 + STEP: get /status 02/25/23 10:46:04.623 + STEP: deleting 02/25/23 10:46:04.632 + STEP: deleting a collection 02/25/23 10:46:04.667 + [AfterEach] [sig-apps] CronJob test/e2e/framework/framework.go:187 - Dec 14 16:34:55.445: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-6337" for this suite. 12/14/22 16:34:55.451 - [AfterEach] [sig-network] Services - test/e2e/network/service.go:762 + Feb 25 10:46:04.709: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "cronjob-3574" for this suite. 02/25/23 10:46:04.72 << End Captured GinkgoWriter Output ------------------------------ -SSSSSS +SSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] Namespaces [Serial] - should ensure that all services are removed when a namespace is deleted [Conformance] - test/e2e/apimachinery/namespace.go:250 -[BeforeEach] [sig-api-machinery] Namespaces [Serial] +[sig-api-machinery] Watchers + should receive events on concurrent watches in same order [Conformance] + test/e2e/apimachinery/watch.go:334 +[BeforeEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:34:55.467 -Dec 14 16:34:55.467: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename namespaces 12/14/22 16:34:55.47 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:55.516 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:55.519 -[It] should ensure that all services are removed when a namespace is deleted [Conformance] - test/e2e/apimachinery/namespace.go:250 -STEP: Creating a test namespace 12/14/22 16:34:55.523 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:55.597 -STEP: Creating a service in the namespace 12/14/22 16:34:55.601 -STEP: Deleting the namespace 12/14/22 16:34:55.614 -STEP: Waiting for the namespace to be removed. 12/14/22 16:34:55.622 -STEP: Recreating the namespace 12/14/22 16:35:01.629 -STEP: Verifying there is no service in the namespace 12/14/22 16:35:01.656 -[AfterEach] [sig-api-machinery] Namespaces [Serial] +STEP: Creating a kubernetes client 02/25/23 10:46:04.736 +Feb 25 10:46:04.737: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename watch 02/25/23 10:46:04.739 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:04.78 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:04.786 +[It] should receive events on concurrent watches in same order [Conformance] + test/e2e/apimachinery/watch.go:334 +STEP: getting a starting resourceVersion 02/25/23 10:46:04.795 +STEP: starting a background goroutine to produce watch events 02/25/23 10:46:04.806 +STEP: creating watches starting from each resource version of the events produced and verifying they all receive resource versions in the same order 02/25/23 10:46:04.806 +[AfterEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:187 -Dec 14 16:35:01.660: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "namespaces-5987" for this suite. 12/14/22 16:35:01.665 -STEP: Destroying namespace "nsdeletetest-8585" for this suite. 12/14/22 16:35:01.674 -Dec 14 16:35:01.678: INFO: Namespace nsdeletetest-8585 was already deleted -STEP: Destroying namespace "nsdeletetest-4140" for this suite. 12/14/22 16:35:01.678 -{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance]","completed":184,"skipped":3382,"failed":0} +Feb 25 10:46:07.534: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "watch-6466" for this suite. 02/25/23 10:46:07.552 +{"msg":"PASSED [sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance]","completed":180,"skipped":3427,"failed":0} ------------------------------ -• [SLOW TEST] [6.219 seconds] -[sig-api-machinery] Namespaces [Serial] +• [2.877 seconds] +[sig-api-machinery] Watchers test/e2e/apimachinery/framework.go:23 - should ensure that all services are removed when a namespace is deleted [Conformance] - test/e2e/apimachinery/namespace.go:250 + should receive events on concurrent watches in same order [Conformance] + test/e2e/apimachinery/watch.go:334 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Namespaces [Serial] + [BeforeEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:34:55.467 - Dec 14 16:34:55.467: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename namespaces 12/14/22 16:34:55.47 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:55.516 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:34:55.519 - [It] should ensure that all services are removed when a namespace is deleted [Conformance] - test/e2e/apimachinery/namespace.go:250 - STEP: Creating a test namespace 12/14/22 16:34:55.523 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:34:55.597 - STEP: Creating a service in the namespace 12/14/22 16:34:55.601 - STEP: Deleting the namespace 12/14/22 16:34:55.614 - STEP: Waiting for the namespace to be removed. 12/14/22 16:34:55.622 - STEP: Recreating the namespace 12/14/22 16:35:01.629 - STEP: Verifying there is no service in the namespace 12/14/22 16:35:01.656 - [AfterEach] [sig-api-machinery] Namespaces [Serial] + STEP: Creating a kubernetes client 02/25/23 10:46:04.736 + Feb 25 10:46:04.737: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename watch 02/25/23 10:46:04.739 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:04.78 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:04.786 + [It] should receive events on concurrent watches in same order [Conformance] + test/e2e/apimachinery/watch.go:334 + STEP: getting a starting resourceVersion 02/25/23 10:46:04.795 + STEP: starting a background goroutine to produce watch events 02/25/23 10:46:04.806 + STEP: creating watches starting from each resource version of the events produced and verifying they all receive resource versions in the same order 02/25/23 10:46:04.806 + [AfterEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:187 - Dec 14 16:35:01.660: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "namespaces-5987" for this suite. 12/14/22 16:35:01.665 - STEP: Destroying namespace "nsdeletetest-8585" for this suite. 12/14/22 16:35:01.674 - Dec 14 16:35:01.678: INFO: Namespace nsdeletetest-8585 was already deleted - STEP: Destroying namespace "nsdeletetest-4140" for this suite. 12/14/22 16:35:01.678 + Feb 25 10:46:07.534: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "watch-6466" for this suite. 02/25/23 10:46:07.552 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSS +S ------------------------------ -[sig-apps] Deployment - deployment should support proportional scaling [Conformance] - test/e2e/apps/deployment.go:160 -[BeforeEach] [sig-apps] Deployment +[sig-node] InitContainer [NodeConformance] + should invoke init containers on a RestartNever pod [Conformance] + test/e2e/common/node/init_container.go:176 +[BeforeEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:35:01.697 -Dec 14 16:35:01.697: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename deployment 12/14/22 16:35:01.698 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:35:01.729 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:35:01.734 -[BeforeEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:91 -[It] deployment should support proportional scaling [Conformance] - test/e2e/apps/deployment.go:160 -Dec 14 16:35:01.738: INFO: Creating deployment "webserver-deployment" -Dec 14 16:35:01.749: INFO: Waiting for observed generation 1 -Dec 14 16:35:03.777: INFO: Waiting for all required pods to come up -Dec 14 16:35:03.786: INFO: Pod name httpd: Found 10 pods out of 10 -STEP: ensuring each pod is running 12/14/22 16:35:03.786 -Dec 14 16:35:03.786: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-5j7qs" in namespace "deployment-5244" to be "running" -Dec 14 16:35:03.787: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-v2g7g" in namespace "deployment-5244" to be "running" -Dec 14 16:35:03.787: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-5jldt" in namespace "deployment-5244" to be "running" -Dec 14 16:35:03.787: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-m8ljs" in namespace "deployment-5244" to be "running" -Dec 14 16:35:03.787: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-pshtw" in namespace "deployment-5244" to be "running" -Dec 14 16:35:03.787: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-r67nb" in namespace "deployment-5244" to be "running" -Dec 14 16:35:03.797: INFO: Pod "webserver-deployment-845c8977d9-m8ljs": Phase="Pending", Reason="", readiness=false. Elapsed: 9.673447ms -Dec 14 16:35:03.797: INFO: Pod "webserver-deployment-845c8977d9-5jldt": Phase="Pending", Reason="", readiness=false. Elapsed: 9.867713ms -Dec 14 16:35:03.800: INFO: Pod "webserver-deployment-845c8977d9-pshtw": Phase="Pending", Reason="", readiness=false. Elapsed: 13.305273ms -Dec 14 16:35:03.801: INFO: Pod "webserver-deployment-845c8977d9-v2g7g": Phase="Pending", Reason="", readiness=false. Elapsed: 13.842356ms -Dec 14 16:35:03.801: INFO: Pod "webserver-deployment-845c8977d9-5j7qs": Phase="Pending", Reason="", readiness=false. Elapsed: 14.588539ms -Dec 14 16:35:03.803: INFO: Pod "webserver-deployment-845c8977d9-r67nb": Phase="Pending", Reason="", readiness=false. Elapsed: 16.023473ms -Dec 14 16:35:05.810: INFO: Pod "webserver-deployment-845c8977d9-m8ljs": Phase="Running", Reason="", readiness=true. Elapsed: 2.022704451s -Dec 14 16:35:05.810: INFO: Pod "webserver-deployment-845c8977d9-m8ljs" satisfied condition "running" -Dec 14 16:35:05.810: INFO: Pod "webserver-deployment-845c8977d9-5jldt": Phase="Running", Reason="", readiness=true. Elapsed: 2.02296123s -Dec 14 16:35:05.810: INFO: Pod "webserver-deployment-845c8977d9-5jldt" satisfied condition "running" -Dec 14 16:35:05.816: INFO: Pod "webserver-deployment-845c8977d9-5j7qs": Phase="Running", Reason="", readiness=true. Elapsed: 2.030041965s -Dec 14 16:35:05.816: INFO: Pod "webserver-deployment-845c8977d9-5j7qs" satisfied condition "running" -Dec 14 16:35:05.817: INFO: Pod "webserver-deployment-845c8977d9-v2g7g": Phase="Running", Reason="", readiness=true. Elapsed: 2.029926412s -Dec 14 16:35:05.817: INFO: Pod "webserver-deployment-845c8977d9-v2g7g" satisfied condition "running" -Dec 14 16:35:05.820: INFO: Pod "webserver-deployment-845c8977d9-r67nb": Phase="Running", Reason="", readiness=true. Elapsed: 2.033005913s -Dec 14 16:35:05.820: INFO: Pod "webserver-deployment-845c8977d9-r67nb" satisfied condition "running" -Dec 14 16:35:05.821: INFO: Pod "webserver-deployment-845c8977d9-pshtw": Phase="Running", Reason="", readiness=true. Elapsed: 2.033496868s -Dec 14 16:35:05.821: INFO: Pod "webserver-deployment-845c8977d9-pshtw" satisfied condition "running" -Dec 14 16:35:05.821: INFO: Waiting for deployment "webserver-deployment" to complete -Dec 14 16:35:05.831: INFO: Updating deployment "webserver-deployment" with a non-existent image -Dec 14 16:35:05.845: INFO: Updating deployment webserver-deployment -Dec 14 16:35:05.845: INFO: Waiting for observed generation 2 -Dec 14 16:35:07.857: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8 -Dec 14 16:35:07.860: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8 -Dec 14 16:35:07.867: INFO: Waiting for the first rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas -Dec 14 16:35:07.882: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0 -Dec 14 16:35:07.882: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5 -Dec 14 16:35:07.902: INFO: Waiting for the second rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas -Dec 14 16:35:07.910: INFO: Verifying that deployment "webserver-deployment" has minimum required number of available replicas -Dec 14 16:35:07.910: INFO: Scaling up the deployment "webserver-deployment" from 10 to 30 -Dec 14 16:35:07.926: INFO: Updating deployment webserver-deployment -Dec 14 16:35:07.926: INFO: Waiting for the replicasets of deployment "webserver-deployment" to have desired number of replicas -Dec 14 16:35:07.940: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20 -Dec 14 16:35:07.945: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13 -[AfterEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:84 -Dec 14 16:35:09.973: INFO: Deployment "webserver-deployment": -&Deployment{ObjectMeta:{webserver-deployment deployment-5244 0ab457f9-1463-4e62-8e36-be80949617bf 20354 3 2022-12-14 16:35:01 +0000 UTC map[name:httpd] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:unavailableReplicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*30,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0027727b8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:33,UpdatedReplicas:13,AvailableReplicas:8,UnavailableReplicas:25,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2022-12-14 16:35:07 +0000 UTC,LastTransitionTime:2022-12-14 16:35:07 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "webserver-deployment-69b7448995" is progressing.,LastUpdateTime:2022-12-14 16:35:08 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,},},ReadyReplicas:8,CollisionCount:nil,},} - -Dec 14 16:35:10.005: INFO: New ReplicaSet "webserver-deployment-69b7448995" of Deployment "webserver-deployment": -&ReplicaSet{ObjectMeta:{webserver-deployment-69b7448995 deployment-5244 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 20339 3 2022-12-14 16:35:05 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment webserver-deployment 0ab457f9-1463-4e62-8e36-be80949617bf 0xc002772c17 0xc002772c18}] [] [{kube-controller-manager Update apps/v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"0ab457f9-1463-4e62-8e36-be80949617bf\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*13,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 69b7448995,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc002772cb8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:13,FullyLabeledReplicas:13,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} -Dec 14 16:35:10.005: INFO: All old ReplicaSets of Deployment "webserver-deployment": -Dec 14 16:35:10.005: INFO: &ReplicaSet{ObjectMeta:{webserver-deployment-845c8977d9 deployment-5244 4cd43053-7095-4f66-9496-24648890aeae 20422 3 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment webserver-deployment 0ab457f9-1463-4e62-8e36-be80949617bf 0xc002772d27 0xc002772d28}] [] [{kube-controller-manager Update apps/v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"0ab457f9-1463-4e62-8e36-be80949617bf\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 16:35:09 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*20,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 845c8977d9,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc002772db8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:20,FullyLabeledReplicas:20,ObservedGeneration:3,ReadyReplicas:9,AvailableReplicas:9,Conditions:[]ReplicaSetCondition{},},} -Dec 14 16:35:10.021: INFO: Pod "webserver-deployment-69b7448995-6cwjs" is not available: -&Pod{ObjectMeta:{webserver-deployment-69b7448995-6cwjs webserver-deployment-69b7448995- deployment-5244 48da0ea5-4851-4975-b869-cf1b8ae8ab32 20328 0 2022-12-14 16:35:07 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc0035dbd57 0xc0035dbd58}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-d9gbb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d9gbb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.025: INFO: Pod "webserver-deployment-69b7448995-8j2fl" is not available: -&Pod{ObjectMeta:{webserver-deployment-69b7448995-8j2fl webserver-deployment-69b7448995- deployment-5244 20e42efb-138e-4dab-9235-8fe69031bf9b 20323 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc0035dbf47 0xc0035dbf48}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-8994v,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8994v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.026: INFO: Pod "webserver-deployment-69b7448995-8nslm" is not available: -&Pod{ObjectMeta:{webserver-deployment-69b7448995-8nslm webserver-deployment-69b7448995- deployment-5244 61d670bf-ad6b-46c4-bd82-bb4eebb448cc 20262 0 2022-12-14 16:35:05 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cba0c0 0xc003cba0c1}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:06 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-cdx2c,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cdx2c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:,StartTime:2022-12-14 16:35:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.026: INFO: Pod "webserver-deployment-69b7448995-cbvmv" is not available: -&Pod{ObjectMeta:{webserver-deployment-69b7448995-cbvmv webserver-deployment-69b7448995- deployment-5244 f924701a-3847-4bff-a2da-1c15ee1cfe2a 20263 0 2022-12-14 16:35:05 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cba2a7 0xc003cba2a8}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:06 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-28jqz,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-28jqz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:,StartTime:2022-12-14 16:35:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.027: INFO: Pod "webserver-deployment-69b7448995-gbm4n" is not available: -&Pod{ObjectMeta:{webserver-deployment-69b7448995-gbm4n webserver-deployment-69b7448995- deployment-5244 f52908e9-2cff-422d-8b73-49a4cb6a411e 20331 0 2022-12-14 16:35:07 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cba497 0xc003cba498}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-ksfjd,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ksfjd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.027: INFO: Pod "webserver-deployment-69b7448995-gz4sd" is not available: -&Pod{ObjectMeta:{webserver-deployment-69b7448995-gz4sd webserver-deployment-69b7448995- deployment-5244 3f3a435a-4e31-4653-9077-ee0ba480b3df 20342 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cba6b7 0xc003cba6b8}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-64pxl,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-64pxl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.027: INFO: Pod "webserver-deployment-69b7448995-h2zrk" is not available: -&Pod{ObjectMeta:{webserver-deployment-69b7448995-h2zrk webserver-deployment-69b7448995- deployment-5244 5537e185-cf43-40d6-aae6-8d1a13d5e72f 20260 0 2022-12-14 16:35:05 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cba830 0xc003cba831}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:06 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-np7b9,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-np7b9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:,StartTime:2022-12-14 16:35:05 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.028: INFO: Pod "webserver-deployment-69b7448995-jpz2v" is not available: -&Pod{ObjectMeta:{webserver-deployment-69b7448995-jpz2v webserver-deployment-69b7448995- deployment-5244 7d487586-10ec-494b-8e56-eb67cf6dd17d 20355 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cbaa37 0xc003cbaa38}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-7k7ks,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7k7ks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.028: INFO: Pod "webserver-deployment-69b7448995-pfcbp" is not available: -&Pod{ObjectMeta:{webserver-deployment-69b7448995-pfcbp webserver-deployment-69b7448995- deployment-5244 e7158dbf-6121-42e2-a61b-df5e300ef330 20261 0 2022-12-14 16:35:05 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cbac37 0xc003cbac38}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:06 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-4gt2m,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4gt2m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 16:35:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.029: INFO: Pod "webserver-deployment-69b7448995-qtxsj" is not available: -&Pod{ObjectMeta:{webserver-deployment-69b7448995-qtxsj webserver-deployment-69b7448995- deployment-5244 a4d4ef63-016b-4712-9f6d-5accc48fb0ab 20376 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cbae37 0xc003cbae38}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-2mc5l,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2mc5l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.029: INFO: Pod "webserver-deployment-69b7448995-t2bgv" is not available: -&Pod{ObjectMeta:{webserver-deployment-69b7448995-t2bgv webserver-deployment-69b7448995- deployment-5244 4ab1d5fe-bdf8-4bed-beff-2b8c73575caa 20333 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cbb027 0xc003cbb028}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-nf5ww,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nf5ww,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.030: INFO: Pod "webserver-deployment-69b7448995-vxbqv" is not available: -&Pod{ObjectMeta:{webserver-deployment-69b7448995-vxbqv webserver-deployment-69b7448995- deployment-5244 b122645a-fe21-465a-a4d6-f66c1c4ca915 20398 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cbb217 0xc003cbb218}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-lbt4r,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lbt4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.030: INFO: Pod "webserver-deployment-69b7448995-wjlfv" is not available: -&Pod{ObjectMeta:{webserver-deployment-69b7448995-wjlfv webserver-deployment-69b7448995- deployment-5244 c90d7349-c1de-4f6a-bd33-526528d2a496 20425 0 2022-12-14 16:35:05 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cbb407 0xc003cbb408}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.111\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-wsxqq,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wsxqq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:10.233.64.111,StartTime:2022-12-14 16:35:05 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ErrImagePull,Message:rpc error: code = Unknown desc = (Mirrors also failed: [mirror.gcr.io/library/webserver:404: reading manifest 404 in mirror.gcr.io/library/webserver: manifest unknown: Failed to fetch "404" from request "/v2/library/webserver/manifests/404".]): docker.io/library/webserver:404: reading manifest 404 in docker.io/library/webserver: errors: -denied: requested access to the resource is denied -unauthorized: authentication required -,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.111,},},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.031: INFO: Pod "webserver-deployment-845c8977d9-4j24m" is not available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-4j24m webserver-deployment-845c8977d9- deployment-5244 c97e9a52-4594-4f22-ab27-cf487184b14c 20340 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc003cbb627 0xc003cbb628}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-8tltc,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8tltc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.031: INFO: Pod "webserver-deployment-845c8977d9-5j7qs" is available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-5j7qs webserver-deployment-845c8977d9- deployment-5244 6498e068-b285-49cc-8a9d-6ea6f81a60c4 20198 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc003cbb790 0xc003cbb791}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:04 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.17\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-r4kc6,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r4kc6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.17,StartTime:2022-12-14 16:35:02 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:03 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://97a4782937e7b67e8c344a54cbd070024f08193b4e9e2e1c15186336f6359a00,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.17,},},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.032: INFO: Pod "webserver-deployment-845c8977d9-6576f" is available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-6576f webserver-deployment-845c8977d9- deployment-5244 37226210-9ef4-4304-8af1-36825e27b60b 20158 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc003cbb977 0xc003cbb978}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:02 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.108\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-v4zmj,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v4zmj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:10.233.64.108,StartTime:2022-12-14 16:35:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://39f7e2f327d9ecca076d5ea274af904a5dcbd7f17aebfed4cec7a9c3566fec15,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.108,},},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.032: INFO: Pod "webserver-deployment-845c8977d9-8r4gf" is not available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-8r4gf webserver-deployment-845c8977d9- deployment-5244 52410aec-8b0b-42c9-aab5-b2b9cd6af69c 20337 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc003cbbb77 0xc003cbbb78}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-47x77,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-47x77,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.033: INFO: Pod "webserver-deployment-845c8977d9-cxpxj" is not available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-cxpxj webserver-deployment-845c8977d9- deployment-5244 acf29036-94a8-461e-a535-1ec1b9d21c01 20346 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc003cbbce0 0xc003cbbce1}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-msqpb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-msqpb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.033: INFO: Pod "webserver-deployment-845c8977d9-dbzx6" is not available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-dbzx6 webserver-deployment-845c8977d9- deployment-5244 26652919-042f-47d6-89e7-48e8ab149780 20413 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc003cbbea7 0xc003cbbea8}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-tkdhk,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tkdhk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.034: INFO: Pod "webserver-deployment-845c8977d9-drmt4" is available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-drmt4 webserver-deployment-845c8977d9- deployment-5244 2c3b4d02-922e-4079-bafc-63b2392c5368 20182 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc00092aa57 0xc00092aa58}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:03 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.109\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-2gwz5,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2gwz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:10.233.65.109,StartTime:2022-12-14 16:35:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://d1abb4fe5f5528fa7e1e6b035c664b4ea1460d1545fa7e6c66159f1cd90e9fa8,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.109,},},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.036: INFO: Pod "webserver-deployment-845c8977d9-dt82g" is not available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-dt82g webserver-deployment-845c8977d9- deployment-5244 83e17c8f-0fd6-4470-b330-7e2c692cb473 20417 0 2022-12-14 16:35:07 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc00092b2e7 0xc00092b2e8}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-bnjl8,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bnjl8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:07 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.038: INFO: Pod "webserver-deployment-845c8977d9-gd8xl" is not available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-gd8xl webserver-deployment-845c8977d9- deployment-5244 8dcd51f1-42a5-4082-8d61-4771322dd828 20392 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc00092b967 0xc00092b968}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-f2lx8,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f2lx8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.038: INFO: Pod "webserver-deployment-845c8977d9-hcldn" is available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-hcldn webserver-deployment-845c8977d9- deployment-5244 beb7ac2f-79df-4785-b9b4-5d746e96c27c 20179 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005ba217 0xc0005ba218}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:03 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.110\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-hk9qr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hk9qr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:10.233.65.110,StartTime:2022-12-14 16:35:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://8c117eeab0549efc8c65d74749e45c830851ed9bc4c5b949fb17828ed1ec583d,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.110,},},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.039: INFO: Pod "webserver-deployment-845c8977d9-m8ljs" is available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-m8ljs webserver-deployment-845c8977d9- deployment-5244 7a22fab3-fb5f-4814-b77a-c92a26784790 20194 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005ba677 0xc0005ba678}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:03 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.109\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-wnbpv,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wnbpv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:10.233.64.109,StartTime:2022-12-14 16:35:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://3bf4ea7b009385af69d79e7e066ebd66417d9b2ad71403dd4073848c944e5e45,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.109,},},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.040: INFO: Pod "webserver-deployment-845c8977d9-ms56p" is available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-ms56p webserver-deployment-845c8977d9- deployment-5244 21869788-b9b1-4b24-8467-e43332ff6aff 20176 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005bab07 0xc0005bab08}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:03 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.108\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-tqk54,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tqk54,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:10.233.65.108,StartTime:2022-12-14 16:35:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://3789c57302b62c35e50e769da2623248432a44c87ad015922290cf8c7055241c,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.108,},},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.040: INFO: Pod "webserver-deployment-845c8977d9-nn6f4" is available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-nn6f4 webserver-deployment-845c8977d9- deployment-5244 14b1266a-0f6d-4b1b-931b-75b5561e1428 20421 0 2022-12-14 16:35:07 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005bafd7 0xc0005bafd8}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.116\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-hr2nc,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hr2nc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:10.233.64.116,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:09 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://a9e3b7a19c1b5db541280bf764e30fb5e14db55a9142a4bb86efeedc33cb6dad,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.116,},},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.044: INFO: Pod "webserver-deployment-845c8977d9-nplkg" is not available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-nplkg webserver-deployment-845c8977d9- deployment-5244 4d894b7d-d2d2-4005-a46e-e0c1334bf701 20312 0 2022-12-14 16:35:07 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005bb1f7 0xc0005bb1f8}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-99nzh,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-99nzh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:07 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 16:35:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.046: INFO: Pod "webserver-deployment-845c8977d9-qssxn" is not available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-qssxn webserver-deployment-845c8977d9- deployment-5244 efb62d02-92d7-401a-9a8c-ff703d62d9ac 20335 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005bb3f7 0xc0005bb3f8}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-9mrp9,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9mrp9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.047: INFO: Pod "webserver-deployment-845c8977d9-r67nb" is available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-r67nb webserver-deployment-845c8977d9- deployment-5244 580f5e27-93cb-4e35-8a3a-e25c05c56f65 20192 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005bb6f0 0xc0005bb6f1}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:03 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.110\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-gbmvp,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gbmvp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:10.233.64.110,StartTime:2022-12-14 16:35:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://221f93913838ebe08ff05b74e087f3f3c5fe45bead9d829d8867293ee93fea2a,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.110,},},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.048: INFO: Pod "webserver-deployment-845c8977d9-rx8nc" is not available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-rx8nc webserver-deployment-845c8977d9- deployment-5244 4d29193d-ba6e-406f-93eb-b5b36975446f 20301 0 2022-12-14 16:35:07 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005bbb37 0xc0005bbb38}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-xdtx4,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xdtx4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.049: INFO: Pod "webserver-deployment-845c8977d9-sfqb6" is not available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-sfqb6 webserver-deployment-845c8977d9- deployment-5244 21f614d6-f740-4525-9b6d-b6b8b64a74ae 20345 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005bbda0 0xc0005bbda1}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-ddz78,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ddz78,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.055: INFO: Pod "webserver-deployment-845c8977d9-v2g7g" is available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-v2g7g webserver-deployment-845c8977d9- deployment-5244 5a3a56c1-2378-4d43-ac95-83eb6a88cde6 20200 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc00344e127 0xc00344e128}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:04 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.19\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-cr8jj,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cr8jj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.19,StartTime:2022-12-14 16:35:02 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:03 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://7fffbda092a2cff4139205a759542a5c213d8739bffcadf027426e98d6882907,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.19,},},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.055: INFO: Pod "webserver-deployment-845c8977d9-vlftv" is not available: -&Pod{ObjectMeta:{webserver-deployment-845c8977d9-vlftv webserver-deployment-845c8977d9- deployment-5244 a5eae1b6-d8ed-454e-81e3-da078b7a1023 20343 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc00344e327 0xc00344e328}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-qkjwf,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qkjwf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -[AfterEach] [sig-apps] Deployment +STEP: Creating a kubernetes client 02/25/23 10:46:07.614 +Feb 25 10:46:07.615: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename init-container 02/25/23 10:46:07.617 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:07.677 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:07.681 +[BeforeEach] [sig-node] InitContainer [NodeConformance] + test/e2e/common/node/init_container.go:164 +[It] should invoke init containers on a RestartNever pod [Conformance] + test/e2e/common/node/init_container.go:176 +STEP: creating the pod 02/25/23 10:46:07.688 +Feb 25 10:46:07.689: INFO: PodSpec: initContainers in spec.initContainers +[AfterEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:187 -Dec 14 16:35:10.056: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "deployment-5244" for this suite. 12/14/22 16:35:10.069 -{"msg":"PASSED [sig-apps] Deployment deployment should support proportional scaling [Conformance]","completed":185,"skipped":3408,"failed":0} +Feb 25 10:46:14.644: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "init-container-928" for this suite. 02/25/23 10:46:14.66 +{"msg":"PASSED [sig-node] InitContainer [NodeConformance] should invoke init containers on a RestartNever pod [Conformance]","completed":181,"skipped":3428,"failed":0} ------------------------------ -• [SLOW TEST] [8.404 seconds] -[sig-apps] Deployment -test/e2e/apps/framework.go:23 - deployment should support proportional scaling [Conformance] - test/e2e/apps/deployment.go:160 +• [SLOW TEST] [7.073 seconds] +[sig-node] InitContainer [NodeConformance] +test/e2e/common/node/framework.go:23 + should invoke init containers on a RestartNever pod [Conformance] + test/e2e/common/node/init_container.go:176 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Deployment + [BeforeEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:35:01.697 - Dec 14 16:35:01.697: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename deployment 12/14/22 16:35:01.698 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:35:01.729 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:35:01.734 - [BeforeEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:91 - [It] deployment should support proportional scaling [Conformance] - test/e2e/apps/deployment.go:160 - Dec 14 16:35:01.738: INFO: Creating deployment "webserver-deployment" - Dec 14 16:35:01.749: INFO: Waiting for observed generation 1 - Dec 14 16:35:03.777: INFO: Waiting for all required pods to come up - Dec 14 16:35:03.786: INFO: Pod name httpd: Found 10 pods out of 10 - STEP: ensuring each pod is running 12/14/22 16:35:03.786 - Dec 14 16:35:03.786: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-5j7qs" in namespace "deployment-5244" to be "running" - Dec 14 16:35:03.787: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-v2g7g" in namespace "deployment-5244" to be "running" - Dec 14 16:35:03.787: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-5jldt" in namespace "deployment-5244" to be "running" - Dec 14 16:35:03.787: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-m8ljs" in namespace "deployment-5244" to be "running" - Dec 14 16:35:03.787: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-pshtw" in namespace "deployment-5244" to be "running" - Dec 14 16:35:03.787: INFO: Waiting up to 5m0s for pod "webserver-deployment-845c8977d9-r67nb" in namespace "deployment-5244" to be "running" - Dec 14 16:35:03.797: INFO: Pod "webserver-deployment-845c8977d9-m8ljs": Phase="Pending", Reason="", readiness=false. Elapsed: 9.673447ms - Dec 14 16:35:03.797: INFO: Pod "webserver-deployment-845c8977d9-5jldt": Phase="Pending", Reason="", readiness=false. Elapsed: 9.867713ms - Dec 14 16:35:03.800: INFO: Pod "webserver-deployment-845c8977d9-pshtw": Phase="Pending", Reason="", readiness=false. Elapsed: 13.305273ms - Dec 14 16:35:03.801: INFO: Pod "webserver-deployment-845c8977d9-v2g7g": Phase="Pending", Reason="", readiness=false. Elapsed: 13.842356ms - Dec 14 16:35:03.801: INFO: Pod "webserver-deployment-845c8977d9-5j7qs": Phase="Pending", Reason="", readiness=false. Elapsed: 14.588539ms - Dec 14 16:35:03.803: INFO: Pod "webserver-deployment-845c8977d9-r67nb": Phase="Pending", Reason="", readiness=false. Elapsed: 16.023473ms - Dec 14 16:35:05.810: INFO: Pod "webserver-deployment-845c8977d9-m8ljs": Phase="Running", Reason="", readiness=true. Elapsed: 2.022704451s - Dec 14 16:35:05.810: INFO: Pod "webserver-deployment-845c8977d9-m8ljs" satisfied condition "running" - Dec 14 16:35:05.810: INFO: Pod "webserver-deployment-845c8977d9-5jldt": Phase="Running", Reason="", readiness=true. Elapsed: 2.02296123s - Dec 14 16:35:05.810: INFO: Pod "webserver-deployment-845c8977d9-5jldt" satisfied condition "running" - Dec 14 16:35:05.816: INFO: Pod "webserver-deployment-845c8977d9-5j7qs": Phase="Running", Reason="", readiness=true. Elapsed: 2.030041965s - Dec 14 16:35:05.816: INFO: Pod "webserver-deployment-845c8977d9-5j7qs" satisfied condition "running" - Dec 14 16:35:05.817: INFO: Pod "webserver-deployment-845c8977d9-v2g7g": Phase="Running", Reason="", readiness=true. Elapsed: 2.029926412s - Dec 14 16:35:05.817: INFO: Pod "webserver-deployment-845c8977d9-v2g7g" satisfied condition "running" - Dec 14 16:35:05.820: INFO: Pod "webserver-deployment-845c8977d9-r67nb": Phase="Running", Reason="", readiness=true. Elapsed: 2.033005913s - Dec 14 16:35:05.820: INFO: Pod "webserver-deployment-845c8977d9-r67nb" satisfied condition "running" - Dec 14 16:35:05.821: INFO: Pod "webserver-deployment-845c8977d9-pshtw": Phase="Running", Reason="", readiness=true. Elapsed: 2.033496868s - Dec 14 16:35:05.821: INFO: Pod "webserver-deployment-845c8977d9-pshtw" satisfied condition "running" - Dec 14 16:35:05.821: INFO: Waiting for deployment "webserver-deployment" to complete - Dec 14 16:35:05.831: INFO: Updating deployment "webserver-deployment" with a non-existent image - Dec 14 16:35:05.845: INFO: Updating deployment webserver-deployment - Dec 14 16:35:05.845: INFO: Waiting for observed generation 2 - Dec 14 16:35:07.857: INFO: Waiting for the first rollout's replicaset to have .status.availableReplicas = 8 - Dec 14 16:35:07.860: INFO: Waiting for the first rollout's replicaset to have .spec.replicas = 8 - Dec 14 16:35:07.867: INFO: Waiting for the first rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas - Dec 14 16:35:07.882: INFO: Verifying that the second rollout's replicaset has .status.availableReplicas = 0 - Dec 14 16:35:07.882: INFO: Waiting for the second rollout's replicaset to have .spec.replicas = 5 - Dec 14 16:35:07.902: INFO: Waiting for the second rollout's replicaset of deployment "webserver-deployment" to have desired number of replicas - Dec 14 16:35:07.910: INFO: Verifying that deployment "webserver-deployment" has minimum required number of available replicas - Dec 14 16:35:07.910: INFO: Scaling up the deployment "webserver-deployment" from 10 to 30 - Dec 14 16:35:07.926: INFO: Updating deployment webserver-deployment - Dec 14 16:35:07.926: INFO: Waiting for the replicasets of deployment "webserver-deployment" to have desired number of replicas - Dec 14 16:35:07.940: INFO: Verifying that first rollout's replicaset has .spec.replicas = 20 - Dec 14 16:35:07.945: INFO: Verifying that second rollout's replicaset has .spec.replicas = 13 - [AfterEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:84 - Dec 14 16:35:09.973: INFO: Deployment "webserver-deployment": - &Deployment{ObjectMeta:{webserver-deployment deployment-5244 0ab457f9-1463-4e62-8e36-be80949617bf 20354 3 2022-12-14 16:35:01 +0000 UTC map[name:httpd] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:unavailableReplicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*30,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0027727b8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:2,MaxSurge:3,},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:3,Replicas:33,UpdatedReplicas:13,AvailableReplicas:8,UnavailableReplicas:25,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2022-12-14 16:35:07 +0000 UTC,LastTransitionTime:2022-12-14 16:35:07 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "webserver-deployment-69b7448995" is progressing.,LastUpdateTime:2022-12-14 16:35:08 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,},},ReadyReplicas:8,CollisionCount:nil,},} - - Dec 14 16:35:10.005: INFO: New ReplicaSet "webserver-deployment-69b7448995" of Deployment "webserver-deployment": - &ReplicaSet{ObjectMeta:{webserver-deployment-69b7448995 deployment-5244 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 20339 3 2022-12-14 16:35:05 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment webserver-deployment 0ab457f9-1463-4e62-8e36-be80949617bf 0xc002772c17 0xc002772c18}] [] [{kube-controller-manager Update apps/v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"0ab457f9-1463-4e62-8e36-be80949617bf\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*13,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 69b7448995,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [] [] []} {[] [] [{httpd webserver:404 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc002772cb8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:13,FullyLabeledReplicas:13,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} - Dec 14 16:35:10.005: INFO: All old ReplicaSets of Deployment "webserver-deployment": - Dec 14 16:35:10.005: INFO: &ReplicaSet{ObjectMeta:{webserver-deployment-845c8977d9 deployment-5244 4cd43053-7095-4f66-9496-24648890aeae 20422 3 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[deployment.kubernetes.io/desired-replicas:30 deployment.kubernetes.io/max-replicas:33 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment webserver-deployment 0ab457f9-1463-4e62-8e36-be80949617bf 0xc002772d27 0xc002772d28}] [] [{kube-controller-manager Update apps/v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"0ab457f9-1463-4e62-8e36-be80949617bf\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 16:35:09 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*20,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 845c8977d9,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc002772db8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:20,FullyLabeledReplicas:20,ObservedGeneration:3,ReadyReplicas:9,AvailableReplicas:9,Conditions:[]ReplicaSetCondition{},},} - Dec 14 16:35:10.021: INFO: Pod "webserver-deployment-69b7448995-6cwjs" is not available: - &Pod{ObjectMeta:{webserver-deployment-69b7448995-6cwjs webserver-deployment-69b7448995- deployment-5244 48da0ea5-4851-4975-b869-cf1b8ae8ab32 20328 0 2022-12-14 16:35:07 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc0035dbd57 0xc0035dbd58}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-d9gbb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d9gbb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.025: INFO: Pod "webserver-deployment-69b7448995-8j2fl" is not available: - &Pod{ObjectMeta:{webserver-deployment-69b7448995-8j2fl webserver-deployment-69b7448995- deployment-5244 20e42efb-138e-4dab-9235-8fe69031bf9b 20323 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc0035dbf47 0xc0035dbf48}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-8994v,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8994v,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.026: INFO: Pod "webserver-deployment-69b7448995-8nslm" is not available: - &Pod{ObjectMeta:{webserver-deployment-69b7448995-8nslm webserver-deployment-69b7448995- deployment-5244 61d670bf-ad6b-46c4-bd82-bb4eebb448cc 20262 0 2022-12-14 16:35:05 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cba0c0 0xc003cba0c1}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:06 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-cdx2c,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cdx2c,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:,StartTime:2022-12-14 16:35:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.026: INFO: Pod "webserver-deployment-69b7448995-cbvmv" is not available: - &Pod{ObjectMeta:{webserver-deployment-69b7448995-cbvmv webserver-deployment-69b7448995- deployment-5244 f924701a-3847-4bff-a2da-1c15ee1cfe2a 20263 0 2022-12-14 16:35:05 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cba2a7 0xc003cba2a8}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:06 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-28jqz,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-28jqz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:,StartTime:2022-12-14 16:35:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.027: INFO: Pod "webserver-deployment-69b7448995-gbm4n" is not available: - &Pod{ObjectMeta:{webserver-deployment-69b7448995-gbm4n webserver-deployment-69b7448995- deployment-5244 f52908e9-2cff-422d-8b73-49a4cb6a411e 20331 0 2022-12-14 16:35:07 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cba497 0xc003cba498}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-ksfjd,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ksfjd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.027: INFO: Pod "webserver-deployment-69b7448995-gz4sd" is not available: - &Pod{ObjectMeta:{webserver-deployment-69b7448995-gz4sd webserver-deployment-69b7448995- deployment-5244 3f3a435a-4e31-4653-9077-ee0ba480b3df 20342 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cba6b7 0xc003cba6b8}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-64pxl,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-64pxl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.027: INFO: Pod "webserver-deployment-69b7448995-h2zrk" is not available: - &Pod{ObjectMeta:{webserver-deployment-69b7448995-h2zrk webserver-deployment-69b7448995- deployment-5244 5537e185-cf43-40d6-aae6-8d1a13d5e72f 20260 0 2022-12-14 16:35:05 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cba830 0xc003cba831}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:06 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-np7b9,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-np7b9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:,StartTime:2022-12-14 16:35:05 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.028: INFO: Pod "webserver-deployment-69b7448995-jpz2v" is not available: - &Pod{ObjectMeta:{webserver-deployment-69b7448995-jpz2v webserver-deployment-69b7448995- deployment-5244 7d487586-10ec-494b-8e56-eb67cf6dd17d 20355 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cbaa37 0xc003cbaa38}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-7k7ks,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-7k7ks,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.028: INFO: Pod "webserver-deployment-69b7448995-pfcbp" is not available: - &Pod{ObjectMeta:{webserver-deployment-69b7448995-pfcbp webserver-deployment-69b7448995- deployment-5244 e7158dbf-6121-42e2-a61b-df5e300ef330 20261 0 2022-12-14 16:35:05 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cbac37 0xc003cbac38}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:06 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-4gt2m,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4gt2m,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:06 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 16:35:06 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.029: INFO: Pod "webserver-deployment-69b7448995-qtxsj" is not available: - &Pod{ObjectMeta:{webserver-deployment-69b7448995-qtxsj webserver-deployment-69b7448995- deployment-5244 a4d4ef63-016b-4712-9f6d-5accc48fb0ab 20376 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cbae37 0xc003cbae38}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-2mc5l,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2mc5l,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.029: INFO: Pod "webserver-deployment-69b7448995-t2bgv" is not available: - &Pod{ObjectMeta:{webserver-deployment-69b7448995-t2bgv webserver-deployment-69b7448995- deployment-5244 4ab1d5fe-bdf8-4bed-beff-2b8c73575caa 20333 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cbb027 0xc003cbb028}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-nf5ww,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nf5ww,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.030: INFO: Pod "webserver-deployment-69b7448995-vxbqv" is not available: - &Pod{ObjectMeta:{webserver-deployment-69b7448995-vxbqv webserver-deployment-69b7448995- deployment-5244 b122645a-fe21-465a-a4d6-f66c1c4ca915 20398 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cbb217 0xc003cbb218}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-lbt4r,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lbt4r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.030: INFO: Pod "webserver-deployment-69b7448995-wjlfv" is not available: - &Pod{ObjectMeta:{webserver-deployment-69b7448995-wjlfv webserver-deployment-69b7448995- deployment-5244 c90d7349-c1de-4f6a-bd33-526528d2a496 20425 0 2022-12-14 16:35:05 +0000 UTC map[name:httpd pod-template-hash:69b7448995] map[] [{apps/v1 ReplicaSet webserver-deployment-69b7448995 1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae 0xc003cbb407 0xc003cbb408}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:05 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"1bc9e0fd-84fd-4cea-b9ec-6d09f9f578ae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.111\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-wsxqq,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:webserver:404,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wsxqq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:05 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:10.233.64.111,StartTime:2022-12-14 16:35:05 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ErrImagePull,Message:rpc error: code = Unknown desc = (Mirrors also failed: [mirror.gcr.io/library/webserver:404: reading manifest 404 in mirror.gcr.io/library/webserver: manifest unknown: Failed to fetch "404" from request "/v2/library/webserver/manifests/404".]): docker.io/library/webserver:404: reading manifest 404 in docker.io/library/webserver: errors: - denied: requested access to the resource is denied - unauthorized: authentication required - ,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:webserver:404,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.111,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.031: INFO: Pod "webserver-deployment-845c8977d9-4j24m" is not available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-4j24m webserver-deployment-845c8977d9- deployment-5244 c97e9a52-4594-4f22-ab27-cf487184b14c 20340 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc003cbb627 0xc003cbb628}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-8tltc,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8tltc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.031: INFO: Pod "webserver-deployment-845c8977d9-5j7qs" is available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-5j7qs webserver-deployment-845c8977d9- deployment-5244 6498e068-b285-49cc-8a9d-6ea6f81a60c4 20198 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc003cbb790 0xc003cbb791}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:04 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.17\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-r4kc6,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-r4kc6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.17,StartTime:2022-12-14 16:35:02 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:03 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://97a4782937e7b67e8c344a54cbd070024f08193b4e9e2e1c15186336f6359a00,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.17,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.032: INFO: Pod "webserver-deployment-845c8977d9-6576f" is available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-6576f webserver-deployment-845c8977d9- deployment-5244 37226210-9ef4-4304-8af1-36825e27b60b 20158 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc003cbb977 0xc003cbb978}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:02 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.108\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-v4zmj,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v4zmj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:10.233.64.108,StartTime:2022-12-14 16:35:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://39f7e2f327d9ecca076d5ea274af904a5dcbd7f17aebfed4cec7a9c3566fec15,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.108,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.032: INFO: Pod "webserver-deployment-845c8977d9-8r4gf" is not available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-8r4gf webserver-deployment-845c8977d9- deployment-5244 52410aec-8b0b-42c9-aab5-b2b9cd6af69c 20337 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc003cbbb77 0xc003cbbb78}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-47x77,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-47x77,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.033: INFO: Pod "webserver-deployment-845c8977d9-cxpxj" is not available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-cxpxj webserver-deployment-845c8977d9- deployment-5244 acf29036-94a8-461e-a535-1ec1b9d21c01 20346 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc003cbbce0 0xc003cbbce1}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-msqpb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-msqpb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.033: INFO: Pod "webserver-deployment-845c8977d9-dbzx6" is not available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-dbzx6 webserver-deployment-845c8977d9- deployment-5244 26652919-042f-47d6-89e7-48e8ab149780 20413 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc003cbbea7 0xc003cbbea8}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-tkdhk,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tkdhk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.034: INFO: Pod "webserver-deployment-845c8977d9-drmt4" is available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-drmt4 webserver-deployment-845c8977d9- deployment-5244 2c3b4d02-922e-4079-bafc-63b2392c5368 20182 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc00092aa57 0xc00092aa58}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:03 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.109\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-2gwz5,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2gwz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:10.233.65.109,StartTime:2022-12-14 16:35:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://d1abb4fe5f5528fa7e1e6b035c664b4ea1460d1545fa7e6c66159f1cd90e9fa8,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.109,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.036: INFO: Pod "webserver-deployment-845c8977d9-dt82g" is not available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-dt82g webserver-deployment-845c8977d9- deployment-5244 83e17c8f-0fd6-4470-b330-7e2c692cb473 20417 0 2022-12-14 16:35:07 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc00092b2e7 0xc00092b2e8}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-bnjl8,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bnjl8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:07 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.038: INFO: Pod "webserver-deployment-845c8977d9-gd8xl" is not available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-gd8xl webserver-deployment-845c8977d9- deployment-5244 8dcd51f1-42a5-4082-8d61-4771322dd828 20392 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc00092b967 0xc00092b968}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-f2lx8,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f2lx8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.038: INFO: Pod "webserver-deployment-845c8977d9-hcldn" is available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-hcldn webserver-deployment-845c8977d9- deployment-5244 beb7ac2f-79df-4785-b9b4-5d746e96c27c 20179 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005ba217 0xc0005ba218}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:03 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.110\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-hk9qr,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hk9qr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:10.233.65.110,StartTime:2022-12-14 16:35:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://8c117eeab0549efc8c65d74749e45c830851ed9bc4c5b949fb17828ed1ec583d,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.110,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.039: INFO: Pod "webserver-deployment-845c8977d9-m8ljs" is available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-m8ljs webserver-deployment-845c8977d9- deployment-5244 7a22fab3-fb5f-4814-b77a-c92a26784790 20194 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005ba677 0xc0005ba678}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:03 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.109\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-wnbpv,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wnbpv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:10.233.64.109,StartTime:2022-12-14 16:35:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://3bf4ea7b009385af69d79e7e066ebd66417d9b2ad71403dd4073848c944e5e45,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.109,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.040: INFO: Pod "webserver-deployment-845c8977d9-ms56p" is available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-ms56p webserver-deployment-845c8977d9- deployment-5244 21869788-b9b1-4b24-8467-e43332ff6aff 20176 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005bab07 0xc0005bab08}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:03 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.108\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-tqk54,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tqk54,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:10.233.65.108,StartTime:2022-12-14 16:35:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://3789c57302b62c35e50e769da2623248432a44c87ad015922290cf8c7055241c,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.108,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.040: INFO: Pod "webserver-deployment-845c8977d9-nn6f4" is available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-nn6f4 webserver-deployment-845c8977d9- deployment-5244 14b1266a-0f6d-4b1b-931b-75b5561e1428 20421 0 2022-12-14 16:35:07 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005bafd7 0xc0005bafd8}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:09 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.116\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-hr2nc,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hr2nc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:09 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:10.233.64.116,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:09 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://a9e3b7a19c1b5db541280bf764e30fb5e14db55a9142a4bb86efeedc33cb6dad,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.116,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.044: INFO: Pod "webserver-deployment-845c8977d9-nplkg" is not available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-nplkg webserver-deployment-845c8977d9- deployment-5244 4d894b7d-d2d2-4005-a46e-e0c1334bf701 20312 0 2022-12-14 16:35:07 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005bb1f7 0xc0005bb1f8}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-99nzh,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-99nzh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:07 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:07 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:07 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 16:35:07 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.046: INFO: Pod "webserver-deployment-845c8977d9-qssxn" is not available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-qssxn webserver-deployment-845c8977d9- deployment-5244 efb62d02-92d7-401a-9a8c-ff703d62d9ac 20335 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005bb3f7 0xc0005bb3f8}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-9mrp9,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9mrp9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.047: INFO: Pod "webserver-deployment-845c8977d9-r67nb" is available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-r67nb webserver-deployment-845c8977d9- deployment-5244 580f5e27-93cb-4e35-8a3a-e25c05c56f65 20192 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005bb6f0 0xc0005bb6f1}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:03 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.110\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-gbmvp,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gbmvp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:03 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:10.233.64.110,StartTime:2022-12-14 16:35:01 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:02 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://221f93913838ebe08ff05b74e087f3f3c5fe45bead9d829d8867293ee93fea2a,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.110,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.048: INFO: Pod "webserver-deployment-845c8977d9-rx8nc" is not available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-rx8nc webserver-deployment-845c8977d9- deployment-5244 4d29193d-ba6e-406f-93eb-b5b36975446f 20301 0 2022-12-14 16:35:07 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005bbb37 0xc0005bbb38}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:07 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-xdtx4,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xdtx4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.049: INFO: Pod "webserver-deployment-845c8977d9-sfqb6" is not available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-sfqb6 webserver-deployment-845c8977d9- deployment-5244 21f614d6-f740-4525-9b6d-b6b8b64a74ae 20345 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc0005bbda0 0xc0005bbda1}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-ddz78,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ddz78,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.249,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.055: INFO: Pod "webserver-deployment-845c8977d9-v2g7g" is available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-v2g7g webserver-deployment-845c8977d9- deployment-5244 5a3a56c1-2378-4d43-ac95-83eb6a88cde6 20200 0 2022-12-14 16:35:01 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc00344e127 0xc00344e128}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:01 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:04 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.19\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-cr8jj,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cr8jj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:02 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:04 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:01 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.19,StartTime:2022-12-14 16:35:02 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 16:35:03 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://7fffbda092a2cff4139205a759542a5c213d8739bffcadf027426e98d6882907,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.19,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.055: INFO: Pod "webserver-deployment-845c8977d9-vlftv" is not available: - &Pod{ObjectMeta:{webserver-deployment-845c8977d9-vlftv webserver-deployment-845c8977d9- deployment-5244 a5eae1b6-d8ed-454e-81e3-da078b7a1023 20343 0 2022-12-14 16:35:08 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet webserver-deployment-845c8977d9 4cd43053-7095-4f66-9496-24648890aeae 0xc00344e327 0xc00344e328}] [] [{kube-controller-manager Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"4cd43053-7095-4f66-9496-24648890aeae\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:35:08 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-qkjwf,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qkjwf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:35:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 16:35:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - [AfterEach] [sig-apps] Deployment + STEP: Creating a kubernetes client 02/25/23 10:46:07.614 + Feb 25 10:46:07.615: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename init-container 02/25/23 10:46:07.617 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:07.677 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:07.681 + [BeforeEach] [sig-node] InitContainer [NodeConformance] + test/e2e/common/node/init_container.go:164 + [It] should invoke init containers on a RestartNever pod [Conformance] + test/e2e/common/node/init_container.go:176 + STEP: creating the pod 02/25/23 10:46:07.688 + Feb 25 10:46:07.689: INFO: PodSpec: initContainers in spec.initContainers + [AfterEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:187 - Dec 14 16:35:10.056: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "deployment-5244" for this suite. 12/14/22 16:35:10.069 + Feb 25 10:46:14.644: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "init-container-928" for this suite. 02/25/23 10:46:14.66 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSS ------------------------------ -[sig-network] DNS - should support configurable pod DNS nameservers [Conformance] - test/e2e/network/dns.go:411 -[BeforeEach] [sig-network] DNS +[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition + listing custom resource definition objects works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:85 +[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:35:10.111 -Dec 14 16:35:10.111: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename dns 12/14/22 16:35:10.115 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:35:10.155 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:35:10.16 -[It] should support configurable pod DNS nameservers [Conformance] - test/e2e/network/dns.go:411 -STEP: Creating a pod with dnsPolicy=None and customized dnsConfig... 12/14/22 16:35:10.163 -Dec 14 16:35:10.181: INFO: Created pod &Pod{ObjectMeta:{test-dns-nameservers dns-9607 5a6a4658-2828-4b6a-bdf5-bb180f6ece54 20436 0 2022-12-14 16:35:10 +0000 UTC map[] map[] [] [] [{e2e.test Update v1 2022-12-14 16:35:10 +0000 UTC FieldsV1 {"f:spec":{"f:containers":{"k:{\"name\":\"agnhost-container\"}":{".":{},"f:args":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsConfig":{".":{},"f:nameservers":{},"f:searches":{}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-tc7fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost-container,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,Command:[],Args:[pause],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tc7fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:None,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:&PodDNSConfig{Nameservers:[1.1.1.1],Searches:[resolv.conf.local],Options:[]PodDNSConfigOption{},},ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 16:35:10.182: INFO: Waiting up to 5m0s for pod "test-dns-nameservers" in namespace "dns-9607" to be "running and ready" -Dec 14 16:35:10.189: INFO: Pod "test-dns-nameservers": Phase="Pending", Reason="", readiness=false. Elapsed: 7.341361ms -Dec 14 16:35:10.189: INFO: The phase of Pod test-dns-nameservers is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:35:12.200: INFO: Pod "test-dns-nameservers": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018717878s -Dec 14 16:35:12.200: INFO: The phase of Pod test-dns-nameservers is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:35:14.195: INFO: Pod "test-dns-nameservers": Phase="Pending", Reason="", readiness=false. Elapsed: 4.013805039s -Dec 14 16:35:14.195: INFO: The phase of Pod test-dns-nameservers is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:35:16.194: INFO: Pod "test-dns-nameservers": Phase="Running", Reason="", readiness=true. Elapsed: 6.012698766s -Dec 14 16:35:16.194: INFO: The phase of Pod test-dns-nameservers is Running (Ready = true) -Dec 14 16:35:16.194: INFO: Pod "test-dns-nameservers" satisfied condition "running and ready" -STEP: Verifying customized DNS suffix list is configured on pod... 12/14/22 16:35:16.195 -Dec 14 16:35:16.195: INFO: ExecWithOptions {Command:[/agnhost dns-suffix] Namespace:dns-9607 PodName:test-dns-nameservers ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:35:16.195: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:35:16.197: INFO: ExecWithOptions: Clientset creation -Dec 14 16:35:16.197: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/dns-9607/pods/test-dns-nameservers/exec?command=%2Fagnhost&command=dns-suffix&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) -STEP: Verifying customized DNS server is configured on pod... 12/14/22 16:35:16.389 -Dec 14 16:35:16.390: INFO: ExecWithOptions {Command:[/agnhost dns-server-list] Namespace:dns-9607 PodName:test-dns-nameservers ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:35:16.390: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:35:16.393: INFO: ExecWithOptions: Clientset creation -Dec 14 16:35:16.393: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/dns-9607/pods/test-dns-nameservers/exec?command=%2Fagnhost&command=dns-server-list&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) -Dec 14 16:35:16.533: INFO: Deleting pod test-dns-nameservers... -[AfterEach] [sig-network] DNS +STEP: Creating a kubernetes client 02/25/23 10:46:14.689 +Feb 25 10:46:14.690: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename custom-resource-definition 02/25/23 10:46:14.702 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:14.74 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:14.746 +[It] listing custom resource definition objects works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:85 +Feb 25 10:46:14.753: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:35:16.584: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "dns-9607" for this suite. 12/14/22 16:35:16.6 -{"msg":"PASSED [sig-network] DNS should support configurable pod DNS nameservers [Conformance]","completed":186,"skipped":3426,"failed":0} +Feb 25 10:46:22.601: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "custom-resource-definition-3413" for this suite. 02/25/23 10:46:22.61 +{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition listing custom resource definition objects works [Conformance]","completed":182,"skipped":3442,"failed":0} ------------------------------ -• [SLOW TEST] [6.507 seconds] -[sig-network] DNS -test/e2e/network/common/framework.go:23 - should support configurable pod DNS nameservers [Conformance] - test/e2e/network/dns.go:411 +• [SLOW TEST] [7.940 seconds] +[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + Simple CustomResourceDefinition + test/e2e/apimachinery/custom_resource_definition.go:50 + listing custom resource definition objects works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:85 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] DNS + [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:35:10.111 - Dec 14 16:35:10.111: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename dns 12/14/22 16:35:10.115 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:35:10.155 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:35:10.16 - [It] should support configurable pod DNS nameservers [Conformance] - test/e2e/network/dns.go:411 - STEP: Creating a pod with dnsPolicy=None and customized dnsConfig... 12/14/22 16:35:10.163 - Dec 14 16:35:10.181: INFO: Created pod &Pod{ObjectMeta:{test-dns-nameservers dns-9607 5a6a4658-2828-4b6a-bdf5-bb180f6ece54 20436 0 2022-12-14 16:35:10 +0000 UTC map[] map[] [] [] [{e2e.test Update v1 2022-12-14 16:35:10 +0000 UTC FieldsV1 {"f:spec":{"f:containers":{"k:{\"name\":\"agnhost-container\"}":{".":{},"f:args":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsConfig":{".":{},"f:nameservers":{},"f:searches":{}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} }]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-tc7fb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost-container,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,Command:[],Args:[pause],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tc7fb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:None,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:&PodDNSConfig{Nameservers:[1.1.1.1],Searches:[resolv.conf.local],Options:[]PodDNSConfigOption{},},ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{},Message:,Reason:,HostIP:,PodIP:,StartTime:,ContainerStatuses:[]ContainerStatus{},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 16:35:10.182: INFO: Waiting up to 5m0s for pod "test-dns-nameservers" in namespace "dns-9607" to be "running and ready" - Dec 14 16:35:10.189: INFO: Pod "test-dns-nameservers": Phase="Pending", Reason="", readiness=false. Elapsed: 7.341361ms - Dec 14 16:35:10.189: INFO: The phase of Pod test-dns-nameservers is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:35:12.200: INFO: Pod "test-dns-nameservers": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018717878s - Dec 14 16:35:12.200: INFO: The phase of Pod test-dns-nameservers is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:35:14.195: INFO: Pod "test-dns-nameservers": Phase="Pending", Reason="", readiness=false. Elapsed: 4.013805039s - Dec 14 16:35:14.195: INFO: The phase of Pod test-dns-nameservers is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:35:16.194: INFO: Pod "test-dns-nameservers": Phase="Running", Reason="", readiness=true. Elapsed: 6.012698766s - Dec 14 16:35:16.194: INFO: The phase of Pod test-dns-nameservers is Running (Ready = true) - Dec 14 16:35:16.194: INFO: Pod "test-dns-nameservers" satisfied condition "running and ready" - STEP: Verifying customized DNS suffix list is configured on pod... 12/14/22 16:35:16.195 - Dec 14 16:35:16.195: INFO: ExecWithOptions {Command:[/agnhost dns-suffix] Namespace:dns-9607 PodName:test-dns-nameservers ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:35:16.195: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:35:16.197: INFO: ExecWithOptions: Clientset creation - Dec 14 16:35:16.197: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/dns-9607/pods/test-dns-nameservers/exec?command=%2Fagnhost&command=dns-suffix&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) - STEP: Verifying customized DNS server is configured on pod... 12/14/22 16:35:16.389 - Dec 14 16:35:16.390: INFO: ExecWithOptions {Command:[/agnhost dns-server-list] Namespace:dns-9607 PodName:test-dns-nameservers ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:35:16.390: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:35:16.393: INFO: ExecWithOptions: Clientset creation - Dec 14 16:35:16.393: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/dns-9607/pods/test-dns-nameservers/exec?command=%2Fagnhost&command=dns-server-list&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) - Dec 14 16:35:16.533: INFO: Deleting pod test-dns-nameservers... - [AfterEach] [sig-network] DNS + STEP: Creating a kubernetes client 02/25/23 10:46:14.689 + Feb 25 10:46:14.690: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename custom-resource-definition 02/25/23 10:46:14.702 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:14.74 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:14.746 + [It] listing custom resource definition objects works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:85 + Feb 25 10:46:14.753: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:35:16.584: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "dns-9607" for this suite. 12/14/22 16:35:16.6 + Feb 25 10:46:22.601: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "custom-resource-definition-3413" for this suite. 02/25/23 10:46:22.61 << End Captured GinkgoWriter Output ------------------------------ -[sig-storage] Projected configMap - updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:123 -[BeforeEach] [sig-storage] Projected configMap +SSSSSSSSS +------------------------------ +[sig-network] Proxy version v1 + A set of valid responses are returned for both pod and service ProxyWithPath [Conformance] + test/e2e/network/proxy.go:286 +[BeforeEach] version v1 test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:35:16.618 -Dec 14 16:35:16.618: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 16:35:16.623 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:35:16.651 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:35:16.655 -[It] updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:123 -STEP: Creating projection with configMap that has name projected-configmap-test-upd-86a01332-c258-4319-84f5-bed5a35e15da 12/14/22 16:35:16.668 -STEP: Creating the pod 12/14/22 16:35:16.675 -Dec 14 16:35:16.699: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c" in namespace "projected-4312" to be "running and ready" -Dec 14 16:35:16.715: INFO: Pod "pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c": Phase="Pending", Reason="", readiness=false. Elapsed: 15.839945ms -Dec 14 16:35:16.715: INFO: The phase of Pod pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:35:18.722: INFO: Pod "pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023261724s -Dec 14 16:35:18.722: INFO: The phase of Pod pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:35:20.723: INFO: Pod "pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c": Phase="Running", Reason="", readiness=true. Elapsed: 4.024059533s -Dec 14 16:35:20.723: INFO: The phase of Pod pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c is Running (Ready = true) -Dec 14 16:35:20.723: INFO: Pod "pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c" satisfied condition "running and ready" -STEP: Updating configmap projected-configmap-test-upd-86a01332-c258-4319-84f5-bed5a35e15da 12/14/22 16:35:20.744 -STEP: waiting to observe update in volume 12/14/22 16:35:20.751 -[AfterEach] [sig-storage] Projected configMap +STEP: Creating a kubernetes client 02/25/23 10:46:22.635 +Feb 25 10:46:22.635: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename proxy 02/25/23 10:46:22.651 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:22.704 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:22.708 +[It] A set of valid responses are returned for both pod and service ProxyWithPath [Conformance] + test/e2e/network/proxy.go:286 +Feb 25 10:46:22.714: INFO: Creating pod... +Feb 25 10:46:22.734: INFO: Waiting up to 5m0s for pod "agnhost" in namespace "proxy-7051" to be "running" +Feb 25 10:46:22.743: INFO: Pod "agnhost": Phase="Pending", Reason="", readiness=false. Elapsed: 8.116704ms +Feb 25 10:46:24.755: INFO: Pod "agnhost": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020409519s +Feb 25 10:46:26.752: INFO: Pod "agnhost": Phase="Running", Reason="", readiness=true. Elapsed: 4.017819384s +Feb 25 10:46:26.752: INFO: Pod "agnhost" satisfied condition "running" +Feb 25 10:46:26.752: INFO: Creating service... +Feb 25 10:46:26.780: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/pods/agnhost/proxy/some/path/with/DELETE +Feb 25 10:46:26.805: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE +Feb 25 10:46:26.805: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/pods/agnhost/proxy/some/path/with/GET +Feb 25 10:46:26.817: INFO: http.Client request:GET | StatusCode:200 | Response:foo | Method:GET +Feb 25 10:46:26.817: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/pods/agnhost/proxy/some/path/with/HEAD +Feb 25 10:46:26.849: INFO: http.Client request:HEAD | StatusCode:200 +Feb 25 10:46:26.849: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/pods/agnhost/proxy/some/path/with/OPTIONS +Feb 25 10:46:26.857: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS +Feb 25 10:46:26.858: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/pods/agnhost/proxy/some/path/with/PATCH +Feb 25 10:46:26.871: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH +Feb 25 10:46:26.871: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/pods/agnhost/proxy/some/path/with/POST +Feb 25 10:46:26.889: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST +Feb 25 10:46:26.890: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/pods/agnhost/proxy/some/path/with/PUT +Feb 25 10:46:26.901: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT +Feb 25 10:46:26.901: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/services/test-service/proxy/some/path/with/DELETE +Feb 25 10:46:26.946: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE +Feb 25 10:46:26.946: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/services/test-service/proxy/some/path/with/GET +Feb 25 10:46:26.966: INFO: http.Client request:GET | StatusCode:200 | Response:foo | Method:GET +Feb 25 10:46:26.966: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/services/test-service/proxy/some/path/with/HEAD +Feb 25 10:46:26.984: INFO: http.Client request:HEAD | StatusCode:200 +Feb 25 10:46:26.984: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/services/test-service/proxy/some/path/with/OPTIONS +Feb 25 10:46:27.018: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS +Feb 25 10:46:27.018: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/services/test-service/proxy/some/path/with/PATCH +Feb 25 10:46:27.057: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH +Feb 25 10:46:27.057: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/services/test-service/proxy/some/path/with/POST +Feb 25 10:46:27.071: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST +Feb 25 10:46:27.072: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/services/test-service/proxy/some/path/with/PUT +Feb 25 10:46:27.088: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT +[AfterEach] version v1 test/e2e/framework/framework.go:187 -Dec 14 16:35:22.777: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-4312" for this suite. 12/14/22 16:35:22.785 -{"msg":"PASSED [sig-storage] Projected configMap updates should be reflected in volume [NodeConformance] [Conformance]","completed":187,"skipped":3426,"failed":0} +Feb 25 10:46:27.088: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "proxy-7051" for this suite. 02/25/23 10:46:27.119 +{"msg":"PASSED [sig-network] Proxy version v1 A set of valid responses are returned for both pod and service ProxyWithPath [Conformance]","completed":183,"skipped":3451,"failed":0} ------------------------------ -• [SLOW TEST] [6.175 seconds] -[sig-storage] Projected configMap -test/e2e/common/storage/framework.go:23 - updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:123 +• [4.533 seconds] +[sig-network] Proxy +test/e2e/network/common/framework.go:23 + version v1 + test/e2e/network/proxy.go:74 + A set of valid responses are returned for both pod and service ProxyWithPath [Conformance] + test/e2e/network/proxy.go:286 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected configMap + [BeforeEach] version v1 test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:35:16.618 - Dec 14 16:35:16.618: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 16:35:16.623 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:35:16.651 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:35:16.655 - [It] updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:123 - STEP: Creating projection with configMap that has name projected-configmap-test-upd-86a01332-c258-4319-84f5-bed5a35e15da 12/14/22 16:35:16.668 - STEP: Creating the pod 12/14/22 16:35:16.675 - Dec 14 16:35:16.699: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c" in namespace "projected-4312" to be "running and ready" - Dec 14 16:35:16.715: INFO: Pod "pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c": Phase="Pending", Reason="", readiness=false. Elapsed: 15.839945ms - Dec 14 16:35:16.715: INFO: The phase of Pod pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:35:18.722: INFO: Pod "pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023261724s - Dec 14 16:35:18.722: INFO: The phase of Pod pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:35:20.723: INFO: Pod "pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c": Phase="Running", Reason="", readiness=true. Elapsed: 4.024059533s - Dec 14 16:35:20.723: INFO: The phase of Pod pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c is Running (Ready = true) - Dec 14 16:35:20.723: INFO: Pod "pod-projected-configmaps-3a2695d6-c3a3-4de7-8381-11418158763c" satisfied condition "running and ready" - STEP: Updating configmap projected-configmap-test-upd-86a01332-c258-4319-84f5-bed5a35e15da 12/14/22 16:35:20.744 - STEP: waiting to observe update in volume 12/14/22 16:35:20.751 - [AfterEach] [sig-storage] Projected configMap + STEP: Creating a kubernetes client 02/25/23 10:46:22.635 + Feb 25 10:46:22.635: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename proxy 02/25/23 10:46:22.651 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:22.704 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:22.708 + [It] A set of valid responses are returned for both pod and service ProxyWithPath [Conformance] + test/e2e/network/proxy.go:286 + Feb 25 10:46:22.714: INFO: Creating pod... + Feb 25 10:46:22.734: INFO: Waiting up to 5m0s for pod "agnhost" in namespace "proxy-7051" to be "running" + Feb 25 10:46:22.743: INFO: Pod "agnhost": Phase="Pending", Reason="", readiness=false. Elapsed: 8.116704ms + Feb 25 10:46:24.755: INFO: Pod "agnhost": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020409519s + Feb 25 10:46:26.752: INFO: Pod "agnhost": Phase="Running", Reason="", readiness=true. Elapsed: 4.017819384s + Feb 25 10:46:26.752: INFO: Pod "agnhost" satisfied condition "running" + Feb 25 10:46:26.752: INFO: Creating service... + Feb 25 10:46:26.780: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/pods/agnhost/proxy/some/path/with/DELETE + Feb 25 10:46:26.805: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE + Feb 25 10:46:26.805: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/pods/agnhost/proxy/some/path/with/GET + Feb 25 10:46:26.817: INFO: http.Client request:GET | StatusCode:200 | Response:foo | Method:GET + Feb 25 10:46:26.817: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/pods/agnhost/proxy/some/path/with/HEAD + Feb 25 10:46:26.849: INFO: http.Client request:HEAD | StatusCode:200 + Feb 25 10:46:26.849: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/pods/agnhost/proxy/some/path/with/OPTIONS + Feb 25 10:46:26.857: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS + Feb 25 10:46:26.858: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/pods/agnhost/proxy/some/path/with/PATCH + Feb 25 10:46:26.871: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH + Feb 25 10:46:26.871: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/pods/agnhost/proxy/some/path/with/POST + Feb 25 10:46:26.889: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST + Feb 25 10:46:26.890: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/pods/agnhost/proxy/some/path/with/PUT + Feb 25 10:46:26.901: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT + Feb 25 10:46:26.901: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/services/test-service/proxy/some/path/with/DELETE + Feb 25 10:46:26.946: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE + Feb 25 10:46:26.946: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/services/test-service/proxy/some/path/with/GET + Feb 25 10:46:26.966: INFO: http.Client request:GET | StatusCode:200 | Response:foo | Method:GET + Feb 25 10:46:26.966: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/services/test-service/proxy/some/path/with/HEAD + Feb 25 10:46:26.984: INFO: http.Client request:HEAD | StatusCode:200 + Feb 25 10:46:26.984: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/services/test-service/proxy/some/path/with/OPTIONS + Feb 25 10:46:27.018: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS + Feb 25 10:46:27.018: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/services/test-service/proxy/some/path/with/PATCH + Feb 25 10:46:27.057: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH + Feb 25 10:46:27.057: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/services/test-service/proxy/some/path/with/POST + Feb 25 10:46:27.071: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST + Feb 25 10:46:27.072: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-7051/services/test-service/proxy/some/path/with/PUT + Feb 25 10:46:27.088: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT + [AfterEach] version v1 test/e2e/framework/framework.go:187 - Dec 14 16:35:22.777: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-4312" for this suite. 12/14/22 16:35:22.785 + Feb 25 10:46:27.088: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "proxy-7051" for this suite. 02/25/23 10:46:27.119 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - should be able to deny custom resource creation, update and deletion [Conformance] - test/e2e/apimachinery/webhook.go:220 + should mutate custom resource with pruning [Conformance] + test/e2e/apimachinery/webhook.go:340 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:35:22.796 -Dec 14 16:35:22.796: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 16:35:22.798 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:35:22.818 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:35:22.821 +STEP: Creating a kubernetes client 02/25/23 10:46:27.203 +Feb 25 10:46:27.204: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 10:46:27.206 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:27.24 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:27.264 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 16:35:22.841 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:35:23.33 -STEP: Deploying the webhook pod 12/14/22 16:35:23.348 -STEP: Wait for the deployment to be ready 12/14/22 16:35:23.367 -Dec 14 16:35:23.377: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set -Dec 14 16:35:25.394: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:35:27.401: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:35:29.402: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:35:31.403: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:35:33.404: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} -STEP: Deploying the webhook service 12/14/22 16:35:35.403 -STEP: Verifying the service has paired with the endpoint 12/14/22 16:35:35.418 -Dec 14 16:35:36.418: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] should be able to deny custom resource creation, update and deletion [Conformance] - test/e2e/apimachinery/webhook.go:220 -Dec 14 16:35:36.426: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Registering the custom resource webhook via the AdmissionRegistration API 12/14/22 16:35:36.944 -STEP: Creating a custom resource that should be denied by the webhook 12/14/22 16:35:36.975 -STEP: Creating a custom resource whose deletion would be denied by the webhook 12/14/22 16:35:39.112 -STEP: Updating the custom resource with disallowed data should be denied 12/14/22 16:35:39.13 -STEP: Deleting the custom resource should be denied 12/14/22 16:35:39.142 -STEP: Remove the offending key and value from the custom resource data 12/14/22 16:35:39.15 -STEP: Deleting the updated custom resource should be successful 12/14/22 16:35:39.168 +STEP: Setting up server cert 02/25/23 10:46:27.322 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:46:29.888 +STEP: Deploying the webhook pod 02/25/23 10:46:29.904 +STEP: Wait for the deployment to be ready 02/25/23 10:46:29.928 +Feb 25 10:46:29.947: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created +Feb 25 10:46:32.117: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 46, 30, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 46, 30, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 46, 30, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 46, 29, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:46:34.126: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 46, 30, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 46, 30, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 46, 30, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 46, 29, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} +STEP: Deploying the webhook service 02/25/23 10:46:36.125 +STEP: Verifying the service has paired with the endpoint 02/25/23 10:46:36.141 +Feb 25 10:46:37.141: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should mutate custom resource with pruning [Conformance] + test/e2e/apimachinery/webhook.go:340 +Feb 25 10:46:37.148: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Registering the mutating webhook for custom resource e2e-test-webhook-305-crds.webhook.example.com via the AdmissionRegistration API 02/25/23 10:46:37.772 +STEP: Creating a custom resource that should be mutated by the webhook 02/25/23 10:46:37.824 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:35:39.723: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-4953" for this suite. 12/14/22 16:35:39.732 -STEP: Destroying namespace "webhook-4953-markers" for this suite. 12/14/22 16:35:39.745 +Feb 25 10:46:40.793: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-1846" for this suite. 02/25/23 10:46:40.813 +STEP: Destroying namespace "webhook-1846-markers" for this suite. 02/25/23 10:46:40.841 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny custom resource creation, update and deletion [Conformance]","completed":188,"skipped":3451,"failed":0} +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with pruning [Conformance]","completed":184,"skipped":3467,"failed":0} ------------------------------ -• [SLOW TEST] [17.101 seconds] +• [SLOW TEST] [14.263 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - should be able to deny custom resource creation, update and deletion [Conformance] - test/e2e/apimachinery/webhook.go:220 +test/e2e/apimachinery/framework.go:23 + should mutate custom resource with pruning [Conformance] + test/e2e/apimachinery/webhook.go:340 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:35:22.796 - Dec 14 16:35:22.796: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 16:35:22.798 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:35:22.818 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:35:22.821 + STEP: Creating a kubernetes client 02/25/23 10:46:27.203 + Feb 25 10:46:27.204: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 10:46:27.206 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:27.24 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:27.264 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 16:35:22.841 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:35:23.33 - STEP: Deploying the webhook pod 12/14/22 16:35:23.348 - STEP: Wait for the deployment to be ready 12/14/22 16:35:23.367 - Dec 14 16:35:23.377: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set - Dec 14 16:35:25.394: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:35:27.401: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:35:29.402: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:35:31.403: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:35:33.404: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 35, 23, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} - STEP: Deploying the webhook service 12/14/22 16:35:35.403 - STEP: Verifying the service has paired with the endpoint 12/14/22 16:35:35.418 - Dec 14 16:35:36.418: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] should be able to deny custom resource creation, update and deletion [Conformance] - test/e2e/apimachinery/webhook.go:220 - Dec 14 16:35:36.426: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Registering the custom resource webhook via the AdmissionRegistration API 12/14/22 16:35:36.944 - STEP: Creating a custom resource that should be denied by the webhook 12/14/22 16:35:36.975 - STEP: Creating a custom resource whose deletion would be denied by the webhook 12/14/22 16:35:39.112 - STEP: Updating the custom resource with disallowed data should be denied 12/14/22 16:35:39.13 - STEP: Deleting the custom resource should be denied 12/14/22 16:35:39.142 - STEP: Remove the offending key and value from the custom resource data 12/14/22 16:35:39.15 - STEP: Deleting the updated custom resource should be successful 12/14/22 16:35:39.168 + STEP: Setting up server cert 02/25/23 10:46:27.322 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:46:29.888 + STEP: Deploying the webhook pod 02/25/23 10:46:29.904 + STEP: Wait for the deployment to be ready 02/25/23 10:46:29.928 + Feb 25 10:46:29.947: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created + Feb 25 10:46:32.117: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 46, 30, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 46, 30, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 46, 30, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 46, 29, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:46:34.126: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 46, 30, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 46, 30, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 46, 30, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 46, 29, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} + STEP: Deploying the webhook service 02/25/23 10:46:36.125 + STEP: Verifying the service has paired with the endpoint 02/25/23 10:46:36.141 + Feb 25 10:46:37.141: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] should mutate custom resource with pruning [Conformance] + test/e2e/apimachinery/webhook.go:340 + Feb 25 10:46:37.148: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Registering the mutating webhook for custom resource e2e-test-webhook-305-crds.webhook.example.com via the AdmissionRegistration API 02/25/23 10:46:37.772 + STEP: Creating a custom resource that should be mutated by the webhook 02/25/23 10:46:37.824 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:35:39.723: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-4953" for this suite. 12/14/22 16:35:39.732 - STEP: Destroying namespace "webhook-4953-markers" for this suite. 12/14/22 16:35:39.745 + Feb 25 10:46:40.793: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-1846" for this suite. 02/25/23 10:46:40.813 + STEP: Destroying namespace "webhook-1846-markers" for this suite. 02/25/23 10:46:40.841 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:104 << End Captured GinkgoWriter Output ------------------------------ -SSSSSS ------------------------------- -[sig-apps] CronJob - should schedule multiple jobs concurrently [Conformance] - test/e2e/apps/cronjob.go:69 -[BeforeEach] [sig-apps] CronJob - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:35:39.898 -Dec 14 16:35:39.899: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename cronjob 12/14/22 16:35:39.901 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:35:39.959 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:35:39.962 -[It] should schedule multiple jobs concurrently [Conformance] - test/e2e/apps/cronjob.go:69 -STEP: Creating a cronjob 12/14/22 16:35:39.965 -STEP: Ensuring more than one job is running at a time 12/14/22 16:35:39.973 -STEP: Ensuring at least two running jobs exists by listing jobs explicitly 12/14/22 16:37:01.98 -STEP: Removing cronjob 12/14/22 16:37:01.986 -[AfterEach] [sig-apps] CronJob - test/e2e/framework/framework.go:187 -Dec 14 16:37:01.997: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "cronjob-201" for this suite. 12/14/22 16:37:02.005 -{"msg":"PASSED [sig-apps] CronJob should schedule multiple jobs concurrently [Conformance]","completed":189,"skipped":3457,"failed":0} ------------------------------- -• [SLOW TEST] [82.123 seconds] -[sig-apps] CronJob -test/e2e/apps/framework.go:23 - should schedule multiple jobs concurrently [Conformance] - test/e2e/apps/cronjob.go:69 - - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] CronJob - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:35:39.898 - Dec 14 16:35:39.899: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename cronjob 12/14/22 16:35:39.901 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:35:39.959 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:35:39.962 - [It] should schedule multiple jobs concurrently [Conformance] - test/e2e/apps/cronjob.go:69 - STEP: Creating a cronjob 12/14/22 16:35:39.965 - STEP: Ensuring more than one job is running at a time 12/14/22 16:35:39.973 - STEP: Ensuring at least two running jobs exists by listing jobs explicitly 12/14/22 16:37:01.98 - STEP: Removing cronjob 12/14/22 16:37:01.986 - [AfterEach] [sig-apps] CronJob - test/e2e/framework/framework.go:187 - Dec 14 16:37:01.997: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "cronjob-201" for this suite. 12/14/22 16:37:02.005 - << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSSSSSSS +SSSSSSSSSS ------------------------------ -[sig-network] Services - should find a service from listing all namespaces [Conformance] - test/e2e/network/service.go:3206 -[BeforeEach] [sig-network] Services +[sig-node] Pods Extended Pods Set QOS Class + should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] + test/e2e/node/pods.go:161 +[BeforeEach] [sig-node] Pods Extended test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:37:02.027 -Dec 14 16:37:02.028: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 16:37:02.031 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:02.096 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:02.1 -[BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 -[It] should find a service from listing all namespaces [Conformance] - test/e2e/network/service.go:3206 -STEP: fetching services 12/14/22 16:37:02.103 -[AfterEach] [sig-network] Services +STEP: Creating a kubernetes client 02/25/23 10:46:41.478 +Feb 25 10:46:41.478: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename pods 02/25/23 10:46:41.481 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:41.601 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:41.622 +[BeforeEach] Pods Set QOS Class + test/e2e/node/pods.go:152 +[It] should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] + test/e2e/node/pods.go:161 +STEP: creating the pod 02/25/23 10:46:41.633 +STEP: submitting the pod to kubernetes 02/25/23 10:46:41.633 +STEP: verifying QOS class is set on the pod 02/25/23 10:46:41.665 +[AfterEach] [sig-node] Pods Extended test/e2e/framework/framework.go:187 -Dec 14 16:37:02.109: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-3352" for this suite. 12/14/22 16:37:02.125 -[AfterEach] [sig-network] Services - test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should find a service from listing all namespaces [Conformance]","completed":190,"skipped":3470,"failed":0} +Feb 25 10:46:41.682: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-7251" for this suite. 02/25/23 10:46:41.705 +{"msg":"PASSED [sig-node] Pods Extended Pods Set QOS Class should be set on Pods with matching resource requests and limits for memory and cpu [Conformance]","completed":185,"skipped":3477,"failed":0} ------------------------------ -• [0.134 seconds] -[sig-network] Services -test/e2e/network/common/framework.go:23 - should find a service from listing all namespaces [Conformance] - test/e2e/network/service.go:3206 +• [0.276 seconds] +[sig-node] Pods Extended +test/e2e/node/framework.go:23 + Pods Set QOS Class + test/e2e/node/pods.go:150 + should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] + test/e2e/node/pods.go:161 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Services + [BeforeEach] [sig-node] Pods Extended test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:37:02.027 - Dec 14 16:37:02.028: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 16:37:02.031 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:02.096 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:02.1 - [BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 - [It] should find a service from listing all namespaces [Conformance] - test/e2e/network/service.go:3206 - STEP: fetching services 12/14/22 16:37:02.103 - [AfterEach] [sig-network] Services + STEP: Creating a kubernetes client 02/25/23 10:46:41.478 + Feb 25 10:46:41.478: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename pods 02/25/23 10:46:41.481 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:41.601 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:41.622 + [BeforeEach] Pods Set QOS Class + test/e2e/node/pods.go:152 + [It] should be set on Pods with matching resource requests and limits for memory and cpu [Conformance] + test/e2e/node/pods.go:161 + STEP: creating the pod 02/25/23 10:46:41.633 + STEP: submitting the pod to kubernetes 02/25/23 10:46:41.633 + STEP: verifying QOS class is set on the pod 02/25/23 10:46:41.665 + [AfterEach] [sig-node] Pods Extended test/e2e/framework/framework.go:187 - Dec 14 16:37:02.109: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-3352" for this suite. 12/14/22 16:37:02.125 - [AfterEach] [sig-network] Services - test/e2e/network/service.go:762 + Feb 25 10:46:41.682: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "pods-7251" for this suite. 02/25/23 10:46:41.705 << End Captured GinkgoWriter Output ------------------------------ -SSSSSS +SSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] ConfigMap - should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:422 -[BeforeEach] [sig-storage] ConfigMap +[sig-node] Variable Expansion + should allow substituting values in a container's args [NodeConformance] [Conformance] + test/e2e/common/node/expansion.go:91 +[BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:37:02.162 -Dec 14 16:37:02.163: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename configmap 12/14/22 16:37:02.165 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:02.218 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:02.222 -[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:422 -STEP: Creating configMap with name configmap-test-volume-4d31d62b-07a2-476f-9ace-c3ad372cfe43 12/14/22 16:37:02.225 -STEP: Creating a pod to test consume configMaps 12/14/22 16:37:02.233 -Dec 14 16:37:02.245: INFO: Waiting up to 5m0s for pod "pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1" in namespace "configmap-4727" to be "Succeeded or Failed" -Dec 14 16:37:02.250: INFO: Pod "pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1": Phase="Pending", Reason="", readiness=false. Elapsed: 4.977146ms -Dec 14 16:37:04.257: INFO: Pod "pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011748258s -Dec 14 16:37:06.258: INFO: Pod "pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013172037s -STEP: Saw pod success 12/14/22 16:37:06.258 -Dec 14 16:37:06.259: INFO: Pod "pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1" satisfied condition "Succeeded or Failed" -Dec 14 16:37:06.265: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1 container configmap-volume-test: -STEP: delete the pod 12/14/22 16:37:06.286 -Dec 14 16:37:06.307: INFO: Waiting for pod pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1 to disappear -Dec 14 16:37:06.313: INFO: Pod pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1 no longer exists -[AfterEach] [sig-storage] ConfigMap +STEP: Creating a kubernetes client 02/25/23 10:46:41.757 +Feb 25 10:46:41.757: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename var-expansion 02/25/23 10:46:41.761 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:41.885 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:41.891 +[It] should allow substituting values in a container's args [NodeConformance] [Conformance] + test/e2e/common/node/expansion.go:91 +STEP: Creating a pod to test substitution in container's args 02/25/23 10:46:41.899 +Feb 25 10:46:41.938: INFO: Waiting up to 5m0s for pod "var-expansion-466f0971-97c0-4ec2-841a-36790d303a15" in namespace "var-expansion-1952" to be "Succeeded or Failed" +Feb 25 10:46:41.953: INFO: Pod "var-expansion-466f0971-97c0-4ec2-841a-36790d303a15": Phase="Pending", Reason="", readiness=false. Elapsed: 14.93694ms +Feb 25 10:46:43.962: INFO: Pod "var-expansion-466f0971-97c0-4ec2-841a-36790d303a15": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024213352s +Feb 25 10:46:45.961: INFO: Pod "var-expansion-466f0971-97c0-4ec2-841a-36790d303a15": Phase="Pending", Reason="", readiness=false. Elapsed: 4.023521774s +Feb 25 10:46:47.965: INFO: Pod "var-expansion-466f0971-97c0-4ec2-841a-36790d303a15": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.02700706s +STEP: Saw pod success 02/25/23 10:46:47.965 +Feb 25 10:46:47.965: INFO: Pod "var-expansion-466f0971-97c0-4ec2-841a-36790d303a15" satisfied condition "Succeeded or Failed" +Feb 25 10:46:47.973: INFO: Trying to get logs from node diek3dua9gay-3 pod var-expansion-466f0971-97c0-4ec2-841a-36790d303a15 container dapi-container: +STEP: delete the pod 02/25/23 10:46:48.072 +Feb 25 10:46:48.103: INFO: Waiting for pod var-expansion-466f0971-97c0-4ec2-841a-36790d303a15 to disappear +Feb 25 10:46:48.115: INFO: Pod var-expansion-466f0971-97c0-4ec2-841a-36790d303a15 no longer exists +[AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 -Dec 14 16:37:06.314: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-4727" for this suite. 12/14/22 16:37:06.32 -{"msg":"PASSED [sig-storage] ConfigMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","completed":191,"skipped":3476,"failed":0} +Feb 25 10:46:48.117: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "var-expansion-1952" for this suite. 02/25/23 10:46:48.129 +{"msg":"PASSED [sig-node] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance]","completed":186,"skipped":3496,"failed":0} ------------------------------ -• [4.168 seconds] -[sig-storage] ConfigMap -test/e2e/common/storage/framework.go:23 - should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:422 +• [SLOW TEST] [6.396 seconds] +[sig-node] Variable Expansion +test/e2e/common/node/framework.go:23 + should allow substituting values in a container's args [NodeConformance] [Conformance] + test/e2e/common/node/expansion.go:91 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] ConfigMap + [BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:37:02.162 - Dec 14 16:37:02.163: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename configmap 12/14/22 16:37:02.165 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:02.218 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:02.222 - [It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:422 - STEP: Creating configMap with name configmap-test-volume-4d31d62b-07a2-476f-9ace-c3ad372cfe43 12/14/22 16:37:02.225 - STEP: Creating a pod to test consume configMaps 12/14/22 16:37:02.233 - Dec 14 16:37:02.245: INFO: Waiting up to 5m0s for pod "pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1" in namespace "configmap-4727" to be "Succeeded or Failed" - Dec 14 16:37:02.250: INFO: Pod "pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1": Phase="Pending", Reason="", readiness=false. Elapsed: 4.977146ms - Dec 14 16:37:04.257: INFO: Pod "pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011748258s - Dec 14 16:37:06.258: INFO: Pod "pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013172037s - STEP: Saw pod success 12/14/22 16:37:06.258 - Dec 14 16:37:06.259: INFO: Pod "pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1" satisfied condition "Succeeded or Failed" - Dec 14 16:37:06.265: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1 container configmap-volume-test: - STEP: delete the pod 12/14/22 16:37:06.286 - Dec 14 16:37:06.307: INFO: Waiting for pod pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1 to disappear - Dec 14 16:37:06.313: INFO: Pod pod-configmaps-e6d9aece-c4d0-435a-b81f-2688b56596c1 no longer exists - [AfterEach] [sig-storage] ConfigMap + STEP: Creating a kubernetes client 02/25/23 10:46:41.757 + Feb 25 10:46:41.757: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename var-expansion 02/25/23 10:46:41.761 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:41.885 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:41.891 + [It] should allow substituting values in a container's args [NodeConformance] [Conformance] + test/e2e/common/node/expansion.go:91 + STEP: Creating a pod to test substitution in container's args 02/25/23 10:46:41.899 + Feb 25 10:46:41.938: INFO: Waiting up to 5m0s for pod "var-expansion-466f0971-97c0-4ec2-841a-36790d303a15" in namespace "var-expansion-1952" to be "Succeeded or Failed" + Feb 25 10:46:41.953: INFO: Pod "var-expansion-466f0971-97c0-4ec2-841a-36790d303a15": Phase="Pending", Reason="", readiness=false. Elapsed: 14.93694ms + Feb 25 10:46:43.962: INFO: Pod "var-expansion-466f0971-97c0-4ec2-841a-36790d303a15": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024213352s + Feb 25 10:46:45.961: INFO: Pod "var-expansion-466f0971-97c0-4ec2-841a-36790d303a15": Phase="Pending", Reason="", readiness=false. Elapsed: 4.023521774s + Feb 25 10:46:47.965: INFO: Pod "var-expansion-466f0971-97c0-4ec2-841a-36790d303a15": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.02700706s + STEP: Saw pod success 02/25/23 10:46:47.965 + Feb 25 10:46:47.965: INFO: Pod "var-expansion-466f0971-97c0-4ec2-841a-36790d303a15" satisfied condition "Succeeded or Failed" + Feb 25 10:46:47.973: INFO: Trying to get logs from node diek3dua9gay-3 pod var-expansion-466f0971-97c0-4ec2-841a-36790d303a15 container dapi-container: + STEP: delete the pod 02/25/23 10:46:48.072 + Feb 25 10:46:48.103: INFO: Waiting for pod var-expansion-466f0971-97c0-4ec2-841a-36790d303a15 to disappear + Feb 25 10:46:48.115: INFO: Pod var-expansion-466f0971-97c0-4ec2-841a-36790d303a15 no longer exists + [AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 - Dec 14 16:37:06.314: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "configmap-4727" for this suite. 12/14/22 16:37:06.32 + Feb 25 10:46:48.117: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "var-expansion-1952" for this suite. 02/25/23 10:46:48.129 << End Captured GinkgoWriter Output ------------------------------ -SS +SSSSSSSSSSS ------------------------------ -[sig-apps] DisruptionController - should create a PodDisruptionBudget [Conformance] - test/e2e/apps/disruption.go:107 -[BeforeEach] [sig-apps] DisruptionController +[sig-node] PodTemplates + should run the lifecycle of PodTemplates [Conformance] + test/e2e/common/node/podtemplates.go:53 +[BeforeEach] [sig-node] PodTemplates test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:37:06.332 -Dec 14 16:37:06.332: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename disruption 12/14/22 16:37:06.339 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:06.363 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:06.368 -[BeforeEach] [sig-apps] DisruptionController - test/e2e/apps/disruption.go:71 -[It] should create a PodDisruptionBudget [Conformance] - test/e2e/apps/disruption.go:107 -STEP: creating the pdb 12/14/22 16:37:06.372 -STEP: Waiting for the pdb to be processed 12/14/22 16:37:06.38 -STEP: updating the pdb 12/14/22 16:37:08.394 -STEP: Waiting for the pdb to be processed 12/14/22 16:37:08.407 -STEP: patching the pdb 12/14/22 16:37:10.417 -STEP: Waiting for the pdb to be processed 12/14/22 16:37:10.444 -STEP: Waiting for the pdb to be deleted 12/14/22 16:37:12.47 -[AfterEach] [sig-apps] DisruptionController +STEP: Creating a kubernetes client 02/25/23 10:46:48.153 +Feb 25 10:46:48.153: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename podtemplate 02/25/23 10:46:48.158 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:48.204 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:48.211 +[It] should run the lifecycle of PodTemplates [Conformance] + test/e2e/common/node/podtemplates.go:53 +[AfterEach] [sig-node] PodTemplates test/e2e/framework/framework.go:187 -Dec 14 16:37:12.477: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "disruption-7955" for this suite. 12/14/22 16:37:12.483 -{"msg":"PASSED [sig-apps] DisruptionController should create a PodDisruptionBudget [Conformance]","completed":192,"skipped":3478,"failed":0} +Feb 25 10:46:48.293: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "podtemplate-9807" for this suite. 02/25/23 10:46:48.319 +{"msg":"PASSED [sig-node] PodTemplates should run the lifecycle of PodTemplates [Conformance]","completed":187,"skipped":3507,"failed":0} ------------------------------ -• [SLOW TEST] [6.161 seconds] -[sig-apps] DisruptionController -test/e2e/apps/framework.go:23 - should create a PodDisruptionBudget [Conformance] - test/e2e/apps/disruption.go:107 +• [0.184 seconds] +[sig-node] PodTemplates +test/e2e/common/node/framework.go:23 + should run the lifecycle of PodTemplates [Conformance] + test/e2e/common/node/podtemplates.go:53 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] DisruptionController + [BeforeEach] [sig-node] PodTemplates test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:37:06.332 - Dec 14 16:37:06.332: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename disruption 12/14/22 16:37:06.339 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:06.363 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:06.368 - [BeforeEach] [sig-apps] DisruptionController - test/e2e/apps/disruption.go:71 - [It] should create a PodDisruptionBudget [Conformance] - test/e2e/apps/disruption.go:107 - STEP: creating the pdb 12/14/22 16:37:06.372 - STEP: Waiting for the pdb to be processed 12/14/22 16:37:06.38 - STEP: updating the pdb 12/14/22 16:37:08.394 - STEP: Waiting for the pdb to be processed 12/14/22 16:37:08.407 - STEP: patching the pdb 12/14/22 16:37:10.417 - STEP: Waiting for the pdb to be processed 12/14/22 16:37:10.444 - STEP: Waiting for the pdb to be deleted 12/14/22 16:37:12.47 - [AfterEach] [sig-apps] DisruptionController + STEP: Creating a kubernetes client 02/25/23 10:46:48.153 + Feb 25 10:46:48.153: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename podtemplate 02/25/23 10:46:48.158 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:48.204 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:48.211 + [It] should run the lifecycle of PodTemplates [Conformance] + test/e2e/common/node/podtemplates.go:53 + [AfterEach] [sig-node] PodTemplates test/e2e/framework/framework.go:187 - Dec 14 16:37:12.477: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "disruption-7955" for this suite. 12/14/22 16:37:12.483 + Feb 25 10:46:48.293: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "podtemplate-9807" for this suite. 02/25/23 10:46:48.319 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSS +SSSSSS ------------------------------ -[sig-cli] Kubectl client Proxy server - should support proxy with --port 0 [Conformance] - test/e2e/kubectl/kubectl.go:1785 -[BeforeEach] [sig-cli] Kubectl client +[sig-storage] EmptyDir volumes + should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:136 +[BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:37:12.495 -Dec 14 16:37:12.495: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 16:37:12.498 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:12.533 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:12.537 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 -[It] should support proxy with --port 0 [Conformance] - test/e2e/kubectl/kubectl.go:1785 -STEP: starting the proxy server 12/14/22 16:37:12.547 -Dec 14 16:37:12.548: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-1024 proxy -p 0 --disable-filter' -STEP: curling proxy /api/ output 12/14/22 16:37:12.658 -[AfterEach] [sig-cli] Kubectl client +STEP: Creating a kubernetes client 02/25/23 10:46:48.338 +Feb 25 10:46:48.338: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir 02/25/23 10:46:48.342 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:48.375 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:48.383 +[It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:136 +STEP: Creating a pod to test emptydir 0666 on tmpfs 02/25/23 10:46:48.39 +Feb 25 10:46:48.412: INFO: Waiting up to 5m0s for pod "pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf" in namespace "emptydir-6221" to be "Succeeded or Failed" +Feb 25 10:46:48.421: INFO: Pod "pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf": Phase="Pending", Reason="", readiness=false. Elapsed: 9.484002ms +Feb 25 10:46:50.433: INFO: Pod "pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021238402s +Feb 25 10:46:52.431: INFO: Pod "pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf": Phase="Pending", Reason="", readiness=false. Elapsed: 4.018817846s +Feb 25 10:46:54.434: INFO: Pod "pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.021938535s +STEP: Saw pod success 02/25/23 10:46:54.434 +Feb 25 10:46:54.434: INFO: Pod "pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf" satisfied condition "Succeeded or Failed" +Feb 25 10:46:54.443: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf container test-container: +STEP: delete the pod 02/25/23 10:46:54.477 +Feb 25 10:46:54.514: INFO: Waiting for pod pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf to disappear +Feb 25 10:46:54.525: INFO: Pod pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf no longer exists +[AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 -Dec 14 16:37:12.677: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-1024" for this suite. 12/14/22 16:37:12.689 -{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support proxy with --port 0 [Conformance]","completed":193,"skipped":3499,"failed":0} +Feb 25 10:46:54.526: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-6221" for this suite. 02/25/23 10:46:54.539 +{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","completed":188,"skipped":3513,"failed":0} ------------------------------ -• [0.202 seconds] -[sig-cli] Kubectl client -test/e2e/kubectl/framework.go:23 - Proxy server - test/e2e/kubectl/kubectl.go:1778 - should support proxy with --port 0 [Conformance] - test/e2e/kubectl/kubectl.go:1785 +• [SLOW TEST] [6.219 seconds] +[sig-storage] EmptyDir volumes +test/e2e/common/storage/framework.go:23 + should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:136 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-cli] Kubectl client + [BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:37:12.495 - Dec 14 16:37:12.495: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 16:37:12.498 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:12.533 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:12.537 - [BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 - [It] should support proxy with --port 0 [Conformance] - test/e2e/kubectl/kubectl.go:1785 - STEP: starting the proxy server 12/14/22 16:37:12.547 - Dec 14 16:37:12.548: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-1024 proxy -p 0 --disable-filter' - STEP: curling proxy /api/ output 12/14/22 16:37:12.658 - [AfterEach] [sig-cli] Kubectl client + STEP: Creating a kubernetes client 02/25/23 10:46:48.338 + Feb 25 10:46:48.338: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir 02/25/23 10:46:48.342 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:48.375 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:48.383 + [It] should support (non-root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:136 + STEP: Creating a pod to test emptydir 0666 on tmpfs 02/25/23 10:46:48.39 + Feb 25 10:46:48.412: INFO: Waiting up to 5m0s for pod "pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf" in namespace "emptydir-6221" to be "Succeeded or Failed" + Feb 25 10:46:48.421: INFO: Pod "pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf": Phase="Pending", Reason="", readiness=false. Elapsed: 9.484002ms + Feb 25 10:46:50.433: INFO: Pod "pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf": Phase="Pending", Reason="", readiness=false. Elapsed: 2.021238402s + Feb 25 10:46:52.431: INFO: Pod "pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf": Phase="Pending", Reason="", readiness=false. Elapsed: 4.018817846s + Feb 25 10:46:54.434: INFO: Pod "pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.021938535s + STEP: Saw pod success 02/25/23 10:46:54.434 + Feb 25 10:46:54.434: INFO: Pod "pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf" satisfied condition "Succeeded or Failed" + Feb 25 10:46:54.443: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf container test-container: + STEP: delete the pod 02/25/23 10:46:54.477 + Feb 25 10:46:54.514: INFO: Waiting for pod pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf to disappear + Feb 25 10:46:54.525: INFO: Pod pod-82840f5e-38f2-4375-a3a1-ba087e1b5acf no longer exists + [AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 - Dec 14 16:37:12.677: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-1024" for this suite. 12/14/22 16:37:12.689 + Feb 25 10:46:54.526: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-6221" for this suite. 02/25/23 10:46:54.539 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSS +S ------------------------------ [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] - should have a working scale subresource [Conformance] - test/e2e/apps/statefulset.go:846 + Should recreate evicted statefulset [Conformance] + test/e2e/apps/statefulset.go:737 [BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:37:12.699 -Dec 14 16:37:12.699: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename statefulset 12/14/22 16:37:12.7 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:12.723 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:12.727 +STEP: Creating a kubernetes client 02/25/23 10:46:54.56 +Feb 25 10:46:54.562: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename statefulset 02/25/23 10:46:54.565 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:54.605 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:54.613 [BeforeEach] [sig-apps] StatefulSet test/e2e/apps/statefulset.go:96 [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] test/e2e/apps/statefulset.go:111 -STEP: Creating service test in namespace statefulset-6876 12/14/22 16:37:12.732 -[It] should have a working scale subresource [Conformance] - test/e2e/apps/statefulset.go:846 -STEP: Creating statefulset ss in namespace statefulset-6876 12/14/22 16:37:12.74 -Dec 14 16:37:12.758: INFO: Found 0 stateful pods, waiting for 1 -Dec 14 16:37:22.768: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true -STEP: getting scale subresource 12/14/22 16:37:22.78 -STEP: updating a scale subresource 12/14/22 16:37:22.785 -STEP: verifying the statefulset Spec.Replicas was modified 12/14/22 16:37:22.798 -STEP: Patch a scale subresource 12/14/22 16:37:22.805 -STEP: verifying the statefulset Spec.Replicas was modified 12/14/22 16:37:22.815 +STEP: Creating service test in namespace statefulset-7226 02/25/23 10:46:54.619 +[It] Should recreate evicted statefulset [Conformance] + test/e2e/apps/statefulset.go:737 +STEP: Looking for a node to schedule stateful set and pod 02/25/23 10:46:54.634 +STEP: Creating pod with conflicting port in namespace statefulset-7226 02/25/23 10:46:54.643 +STEP: Waiting until pod test-pod will start running in namespace statefulset-7226 02/25/23 10:46:54.663 +Feb 25 10:46:54.663: INFO: Waiting up to 5m0s for pod "test-pod" in namespace "statefulset-7226" to be "running" +Feb 25 10:46:54.672: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 8.539583ms +Feb 25 10:46:56.696: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.032054398s +Feb 25 10:46:58.682: INFO: Pod "test-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.018896921s +Feb 25 10:46:58.682: INFO: Pod "test-pod" satisfied condition "running" +STEP: Creating statefulset with conflicting port in namespace statefulset-7226 02/25/23 10:46:58.683 +STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-7226 02/25/23 10:46:58.7 +Feb 25 10:46:58.771: INFO: Observed stateful pod in namespace: statefulset-7226, name: ss-0, uid: 66b3c6fe-a566-4e20-9d60-ce04dcc49b31, status phase: Pending. Waiting for statefulset controller to delete. +Feb 25 10:46:58.818: INFO: Observed stateful pod in namespace: statefulset-7226, name: ss-0, uid: 66b3c6fe-a566-4e20-9d60-ce04dcc49b31, status phase: Failed. Waiting for statefulset controller to delete. +Feb 25 10:46:58.853: INFO: Observed stateful pod in namespace: statefulset-7226, name: ss-0, uid: 66b3c6fe-a566-4e20-9d60-ce04dcc49b31, status phase: Failed. Waiting for statefulset controller to delete. +Feb 25 10:46:58.863: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-7226 +STEP: Removing pod with conflicting port in namespace statefulset-7226 02/25/23 10:46:58.863 +STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-7226 and will be in running state 02/25/23 10:46:58.949 [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] test/e2e/apps/statefulset.go:122 -Dec 14 16:37:22.824: INFO: Deleting all statefulset in ns statefulset-6876 -Dec 14 16:37:22.844: INFO: Scaling statefulset ss to 0 -Dec 14 16:37:32.884: INFO: Waiting for statefulset status.replicas updated to 0 -Dec 14 16:37:32.890: INFO: Deleting statefulset ss +Feb 25 10:47:02.992: INFO: Deleting all statefulset in ns statefulset-7226 +Feb 25 10:47:03.001: INFO: Scaling statefulset ss to 0 +Feb 25 10:47:13.047: INFO: Waiting for statefulset status.replicas updated to 0 +Feb 25 10:47:13.054: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:187 -Dec 14 16:37:32.908: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "statefulset-6876" for this suite. 12/14/22 16:37:32.915 -{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should have a working scale subresource [Conformance]","completed":194,"skipped":3517,"failed":0} +Feb 25 10:47:13.111: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "statefulset-7226" for this suite. 02/25/23 10:47:13.171 +{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance]","completed":189,"skipped":3514,"failed":0} ------------------------------ -• [SLOW TEST] [20.230 seconds] +• [SLOW TEST] [18.656 seconds] [sig-apps] StatefulSet test/e2e/apps/framework.go:23 Basic StatefulSet functionality [StatefulSetBasic] test/e2e/apps/statefulset.go:101 - should have a working scale subresource [Conformance] - test/e2e/apps/statefulset.go:846 + Should recreate evicted statefulset [Conformance] + test/e2e/apps/statefulset.go:737 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:37:12.699 - Dec 14 16:37:12.699: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename statefulset 12/14/22 16:37:12.7 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:12.723 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:12.727 + STEP: Creating a kubernetes client 02/25/23 10:46:54.56 + Feb 25 10:46:54.562: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename statefulset 02/25/23 10:46:54.565 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:46:54.605 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:46:54.613 [BeforeEach] [sig-apps] StatefulSet test/e2e/apps/statefulset.go:96 [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] test/e2e/apps/statefulset.go:111 - STEP: Creating service test in namespace statefulset-6876 12/14/22 16:37:12.732 - [It] should have a working scale subresource [Conformance] - test/e2e/apps/statefulset.go:846 - STEP: Creating statefulset ss in namespace statefulset-6876 12/14/22 16:37:12.74 - Dec 14 16:37:12.758: INFO: Found 0 stateful pods, waiting for 1 - Dec 14 16:37:22.768: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true - STEP: getting scale subresource 12/14/22 16:37:22.78 - STEP: updating a scale subresource 12/14/22 16:37:22.785 - STEP: verifying the statefulset Spec.Replicas was modified 12/14/22 16:37:22.798 - STEP: Patch a scale subresource 12/14/22 16:37:22.805 - STEP: verifying the statefulset Spec.Replicas was modified 12/14/22 16:37:22.815 + STEP: Creating service test in namespace statefulset-7226 02/25/23 10:46:54.619 + [It] Should recreate evicted statefulset [Conformance] + test/e2e/apps/statefulset.go:737 + STEP: Looking for a node to schedule stateful set and pod 02/25/23 10:46:54.634 + STEP: Creating pod with conflicting port in namespace statefulset-7226 02/25/23 10:46:54.643 + STEP: Waiting until pod test-pod will start running in namespace statefulset-7226 02/25/23 10:46:54.663 + Feb 25 10:46:54.663: INFO: Waiting up to 5m0s for pod "test-pod" in namespace "statefulset-7226" to be "running" + Feb 25 10:46:54.672: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 8.539583ms + Feb 25 10:46:56.696: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.032054398s + Feb 25 10:46:58.682: INFO: Pod "test-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.018896921s + Feb 25 10:46:58.682: INFO: Pod "test-pod" satisfied condition "running" + STEP: Creating statefulset with conflicting port in namespace statefulset-7226 02/25/23 10:46:58.683 + STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-7226 02/25/23 10:46:58.7 + Feb 25 10:46:58.771: INFO: Observed stateful pod in namespace: statefulset-7226, name: ss-0, uid: 66b3c6fe-a566-4e20-9d60-ce04dcc49b31, status phase: Pending. Waiting for statefulset controller to delete. + Feb 25 10:46:58.818: INFO: Observed stateful pod in namespace: statefulset-7226, name: ss-0, uid: 66b3c6fe-a566-4e20-9d60-ce04dcc49b31, status phase: Failed. Waiting for statefulset controller to delete. + Feb 25 10:46:58.853: INFO: Observed stateful pod in namespace: statefulset-7226, name: ss-0, uid: 66b3c6fe-a566-4e20-9d60-ce04dcc49b31, status phase: Failed. Waiting for statefulset controller to delete. + Feb 25 10:46:58.863: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-7226 + STEP: Removing pod with conflicting port in namespace statefulset-7226 02/25/23 10:46:58.863 + STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-7226 and will be in running state 02/25/23 10:46:58.949 [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] test/e2e/apps/statefulset.go:122 - Dec 14 16:37:22.824: INFO: Deleting all statefulset in ns statefulset-6876 - Dec 14 16:37:22.844: INFO: Scaling statefulset ss to 0 - Dec 14 16:37:32.884: INFO: Waiting for statefulset status.replicas updated to 0 - Dec 14 16:37:32.890: INFO: Deleting statefulset ss + Feb 25 10:47:02.992: INFO: Deleting all statefulset in ns statefulset-7226 + Feb 25 10:47:03.001: INFO: Scaling statefulset ss to 0 + Feb 25 10:47:13.047: INFO: Waiting for statefulset status.replicas updated to 0 + Feb 25 10:47:13.054: INFO: Deleting statefulset ss [AfterEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:187 - Dec 14 16:37:32.908: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "statefulset-6876" for this suite. 12/14/22 16:37:32.915 + Feb 25 10:47:13.111: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "statefulset-7226" for this suite. 02/25/23 10:47:13.171 << End Captured GinkgoWriter Output ------------------------------ -SSSSSS +SSS ------------------------------ -[sig-auth] ServiceAccounts - should mount an API token into pods [Conformance] - test/e2e/auth/service_accounts.go:75 -[BeforeEach] [sig-auth] ServiceAccounts +[sig-network] Services + should complete a service status lifecycle [Conformance] + test/e2e/network/service.go:3415 +[BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:37:32.933 -Dec 14 16:37:32.933: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename svcaccounts 12/14/22 16:37:32.935 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:32.959 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:32.962 -[It] should mount an API token into pods [Conformance] - test/e2e/auth/service_accounts.go:75 -Dec 14 16:37:32.988: INFO: Waiting up to 5m0s for pod "pod-service-account-8de78ed8-a20d-451c-a251-47d57f1715dd" in namespace "svcaccounts-1380" to be "running" -Dec 14 16:37:32.998: INFO: Pod "pod-service-account-8de78ed8-a20d-451c-a251-47d57f1715dd": Phase="Pending", Reason="", readiness=false. Elapsed: 10.366424ms -Dec 14 16:37:35.009: INFO: Pod "pod-service-account-8de78ed8-a20d-451c-a251-47d57f1715dd": Phase="Running", Reason="", readiness=true. Elapsed: 2.020918856s -Dec 14 16:37:35.009: INFO: Pod "pod-service-account-8de78ed8-a20d-451c-a251-47d57f1715dd" satisfied condition "running" -STEP: reading a file in the container 12/14/22 16:37:35.009 -Dec 14 16:37:35.010: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-1380 pod-service-account-8de78ed8-a20d-451c-a251-47d57f1715dd -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token' -STEP: reading a file in the container 12/14/22 16:37:35.245 -Dec 14 16:37:35.246: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-1380 pod-service-account-8de78ed8-a20d-451c-a251-47d57f1715dd -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt' -STEP: reading a file in the container 12/14/22 16:37:35.435 -Dec 14 16:37:35.436: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-1380 pod-service-account-8de78ed8-a20d-451c-a251-47d57f1715dd -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace' -Dec 14 16:37:35.678: INFO: Got root ca configmap in namespace "svcaccounts-1380" -[AfterEach] [sig-auth] ServiceAccounts +STEP: Creating a kubernetes client 02/25/23 10:47:13.224 +Feb 25 10:47:13.224: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 10:47:13.23 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:47:13.318 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:47:13.324 +[BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 +[It] should complete a service status lifecycle [Conformance] + test/e2e/network/service.go:3415 +STEP: creating a Service 02/25/23 10:47:13.361 +STEP: watching for the Service to be added 02/25/23 10:47:13.409 +Feb 25 10:47:13.417: INFO: Found Service test-service-npl25 in namespace services-744 with labels: map[test-service-static:true] & ports [{http TCP 80 {0 80 } 0}] +Feb 25 10:47:13.417: INFO: Service test-service-npl25 created +STEP: Getting /status 02/25/23 10:47:13.417 +Feb 25 10:47:13.430: INFO: Service test-service-npl25 has LoadBalancer: {[]} +STEP: patching the ServiceStatus 02/25/23 10:47:13.43 +STEP: watching for the Service to be patched 02/25/23 10:47:13.449 +Feb 25 10:47:13.455: INFO: observed Service test-service-npl25 in namespace services-744 with annotations: map[] & LoadBalancer: {[]} +Feb 25 10:47:13.456: INFO: Found Service test-service-npl25 in namespace services-744 with annotations: map[patchedstatus:true] & LoadBalancer: {[{203.0.113.1 []}]} +Feb 25 10:47:13.456: INFO: Service test-service-npl25 has service status patched +STEP: updating the ServiceStatus 02/25/23 10:47:13.457 +Feb 25 10:47:13.509: INFO: updatedStatus.Conditions: []v1.Condition{v1.Condition{Type:"StatusUpdate", Status:"True", ObservedGeneration:0, LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} +STEP: watching for the Service to be updated 02/25/23 10:47:13.511 +Feb 25 10:47:13.518: INFO: Observed Service test-service-npl25 in namespace services-744 with annotations: map[] & Conditions: {[]} +Feb 25 10:47:13.519: INFO: Observed event: &Service{ObjectMeta:{test-service-npl25 services-744 e749b4b8-2293-43a1-8053-fdf0c2ec5344 30315 0 2023-02-25 10:47:13 +0000 UTC map[test-service-static:true] map[patchedstatus:true] [] [] [{e2e.test Update v1 2023-02-25 10:47:13 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:test-service-static":{}}},"f:spec":{"f:internalTrafficPolicy":{},"f:ports":{".":{},"k:{\"port\":80,\"protocol\":\"TCP\"}":{".":{},"f:name":{},"f:port":{},"f:protocol":{},"f:targetPort":{}}},"f:sessionAffinity":{},"f:type":{}}} } {e2e.test Update v1 2023-02-25 10:47:13 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:patchedstatus":{}}},"f:status":{"f:loadBalancer":{"f:ingress":{}}}} status}]},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:http,Protocol:TCP,Port:80,TargetPort:{0 80 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{},ClusterIP:10.233.3.51,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.233.3.51],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{LoadBalancerIngress{IP:203.0.113.1,Hostname:,Ports:[]PortStatus{},},},},Conditions:[]Condition{},},} +Feb 25 10:47:13.519: INFO: Found Service test-service-npl25 in namespace services-744 with annotations: map[patchedstatus:true] & Conditions: [{StatusUpdate True 0 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] +Feb 25 10:47:13.519: INFO: Service test-service-npl25 has service status updated +STEP: patching the service 02/25/23 10:47:13.519 +STEP: watching for the Service to be patched 02/25/23 10:47:13.546 +Feb 25 10:47:13.550: INFO: observed Service test-service-npl25 in namespace services-744 with labels: map[test-service-static:true] +Feb 25 10:47:13.551: INFO: observed Service test-service-npl25 in namespace services-744 with labels: map[test-service-static:true] +Feb 25 10:47:13.551: INFO: observed Service test-service-npl25 in namespace services-744 with labels: map[test-service-static:true] +Feb 25 10:47:13.558: INFO: Found Service test-service-npl25 in namespace services-744 with labels: map[test-service:patched test-service-static:true] +Feb 25 10:47:13.558: INFO: Service test-service-npl25 patched +STEP: deleting the service 02/25/23 10:47:13.558 +STEP: watching for the Service to be deleted 02/25/23 10:47:13.65 +Feb 25 10:47:13.656: INFO: Observed event: ADDED +Feb 25 10:47:13.662: INFO: Observed event: MODIFIED +Feb 25 10:47:13.664: INFO: Observed event: MODIFIED +Feb 25 10:47:13.665: INFO: Observed event: MODIFIED +Feb 25 10:47:13.666: INFO: Found Service test-service-npl25 in namespace services-744 with labels: map[test-service:patched test-service-static:true] & annotations: map[patchedstatus:true] +Feb 25 10:47:13.667: INFO: Service test-service-npl25 deleted +[AfterEach] [sig-network] Services + test/e2e/framework/framework.go:187 +Feb 25 10:47:13.668: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-744" for this suite. 02/25/23 10:47:13.692 +[AfterEach] [sig-network] Services + test/e2e/network/service.go:762 +{"msg":"PASSED [sig-network] Services should complete a service status lifecycle [Conformance]","completed":190,"skipped":3517,"failed":0} +------------------------------ +• [0.489 seconds] +[sig-network] Services +test/e2e/network/common/framework.go:23 + should complete a service status lifecycle [Conformance] + test/e2e/network/service.go:3415 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-network] Services + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:47:13.224 + Feb 25 10:47:13.224: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 10:47:13.23 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:47:13.318 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:47:13.324 + [BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 + [It] should complete a service status lifecycle [Conformance] + test/e2e/network/service.go:3415 + STEP: creating a Service 02/25/23 10:47:13.361 + STEP: watching for the Service to be added 02/25/23 10:47:13.409 + Feb 25 10:47:13.417: INFO: Found Service test-service-npl25 in namespace services-744 with labels: map[test-service-static:true] & ports [{http TCP 80 {0 80 } 0}] + Feb 25 10:47:13.417: INFO: Service test-service-npl25 created + STEP: Getting /status 02/25/23 10:47:13.417 + Feb 25 10:47:13.430: INFO: Service test-service-npl25 has LoadBalancer: {[]} + STEP: patching the ServiceStatus 02/25/23 10:47:13.43 + STEP: watching for the Service to be patched 02/25/23 10:47:13.449 + Feb 25 10:47:13.455: INFO: observed Service test-service-npl25 in namespace services-744 with annotations: map[] & LoadBalancer: {[]} + Feb 25 10:47:13.456: INFO: Found Service test-service-npl25 in namespace services-744 with annotations: map[patchedstatus:true] & LoadBalancer: {[{203.0.113.1 []}]} + Feb 25 10:47:13.456: INFO: Service test-service-npl25 has service status patched + STEP: updating the ServiceStatus 02/25/23 10:47:13.457 + Feb 25 10:47:13.509: INFO: updatedStatus.Conditions: []v1.Condition{v1.Condition{Type:"StatusUpdate", Status:"True", ObservedGeneration:0, LastTransitionTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), Reason:"E2E", Message:"Set from e2e test"}} + STEP: watching for the Service to be updated 02/25/23 10:47:13.511 + Feb 25 10:47:13.518: INFO: Observed Service test-service-npl25 in namespace services-744 with annotations: map[] & Conditions: {[]} + Feb 25 10:47:13.519: INFO: Observed event: &Service{ObjectMeta:{test-service-npl25 services-744 e749b4b8-2293-43a1-8053-fdf0c2ec5344 30315 0 2023-02-25 10:47:13 +0000 UTC map[test-service-static:true] map[patchedstatus:true] [] [] [{e2e.test Update v1 2023-02-25 10:47:13 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:test-service-static":{}}},"f:spec":{"f:internalTrafficPolicy":{},"f:ports":{".":{},"k:{\"port\":80,\"protocol\":\"TCP\"}":{".":{},"f:name":{},"f:port":{},"f:protocol":{},"f:targetPort":{}}},"f:sessionAffinity":{},"f:type":{}}} } {e2e.test Update v1 2023-02-25 10:47:13 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:patchedstatus":{}}},"f:status":{"f:loadBalancer":{"f:ingress":{}}}} status}]},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:http,Protocol:TCP,Port:80,TargetPort:{0 80 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{},ClusterIP:10.233.3.51,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,LoadBalancerIP:,LoadBalancerSourceRanges:[],ExternalName:,ExternalTrafficPolicy:,HealthCheckNodePort:0,PublishNotReadyAddresses:false,SessionAffinityConfig:nil,IPFamilyPolicy:*SingleStack,ClusterIPs:[10.233.3.51],IPFamilies:[IPv4],AllocateLoadBalancerNodePorts:nil,LoadBalancerClass:nil,InternalTrafficPolicy:*Cluster,},Status:ServiceStatus{LoadBalancer:LoadBalancerStatus{Ingress:[]LoadBalancerIngress{LoadBalancerIngress{IP:203.0.113.1,Hostname:,Ports:[]PortStatus{},},},},Conditions:[]Condition{},},} + Feb 25 10:47:13.519: INFO: Found Service test-service-npl25 in namespace services-744 with annotations: map[patchedstatus:true] & Conditions: [{StatusUpdate True 0 0001-01-01 00:00:00 +0000 UTC E2E Set from e2e test}] + Feb 25 10:47:13.519: INFO: Service test-service-npl25 has service status updated + STEP: patching the service 02/25/23 10:47:13.519 + STEP: watching for the Service to be patched 02/25/23 10:47:13.546 + Feb 25 10:47:13.550: INFO: observed Service test-service-npl25 in namespace services-744 with labels: map[test-service-static:true] + Feb 25 10:47:13.551: INFO: observed Service test-service-npl25 in namespace services-744 with labels: map[test-service-static:true] + Feb 25 10:47:13.551: INFO: observed Service test-service-npl25 in namespace services-744 with labels: map[test-service-static:true] + Feb 25 10:47:13.558: INFO: Found Service test-service-npl25 in namespace services-744 with labels: map[test-service:patched test-service-static:true] + Feb 25 10:47:13.558: INFO: Service test-service-npl25 patched + STEP: deleting the service 02/25/23 10:47:13.558 + STEP: watching for the Service to be deleted 02/25/23 10:47:13.65 + Feb 25 10:47:13.656: INFO: Observed event: ADDED + Feb 25 10:47:13.662: INFO: Observed event: MODIFIED + Feb 25 10:47:13.664: INFO: Observed event: MODIFIED + Feb 25 10:47:13.665: INFO: Observed event: MODIFIED + Feb 25 10:47:13.666: INFO: Found Service test-service-npl25 in namespace services-744 with labels: map[test-service:patched test-service-static:true] & annotations: map[patchedstatus:true] + Feb 25 10:47:13.667: INFO: Service test-service-npl25 deleted + [AfterEach] [sig-network] Services + test/e2e/framework/framework.go:187 + Feb 25 10:47:13.668: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-744" for this suite. 02/25/23 10:47:13.692 + [AfterEach] [sig-network] Services + test/e2e/network/service.go:762 + << End Captured GinkgoWriter Output +------------------------------ +[sig-network] Services + should be able to change the type from NodePort to ExternalName [Conformance] + test/e2e/network/service.go:1523 +[BeforeEach] [sig-network] Services + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:47:13.715 +Feb 25 10:47:13.716: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 10:47:13.719 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:47:13.81 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:47:13.819 +[BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 +[It] should be able to change the type from NodePort to ExternalName [Conformance] + test/e2e/network/service.go:1523 +STEP: creating a service nodeport-service with the type=NodePort in namespace services-8297 02/25/23 10:47:13.836 +STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service 02/25/23 10:47:13.888 +STEP: creating service externalsvc in namespace services-8297 02/25/23 10:47:13.888 +STEP: creating replication controller externalsvc in namespace services-8297 02/25/23 10:47:13.96 +I0225 10:47:13.981982 13 runners.go:193] Created replication controller with name: externalsvc, namespace: services-8297, replica count: 2 +I0225 10:47:17.034902 13 runners.go:193] externalsvc Pods: 2 out of 2 created, 1 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0225 10:47:20.035480 13 runners.go:193] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +STEP: changing the NodePort service to type=ExternalName 02/25/23 10:47:20.044 +Feb 25 10:47:20.095: INFO: Creating new exec pod +Feb 25 10:47:20.125: INFO: Waiting up to 5m0s for pod "execpod9pmzw" in namespace "services-8297" to be "running" +Feb 25 10:47:20.135: INFO: Pod "execpod9pmzw": Phase="Pending", Reason="", readiness=false. Elapsed: 9.623685ms +Feb 25 10:47:22.146: INFO: Pod "execpod9pmzw": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020591862s +Feb 25 10:47:24.146: INFO: Pod "execpod9pmzw": Phase="Running", Reason="", readiness=true. Elapsed: 4.020382222s +Feb 25 10:47:24.146: INFO: Pod "execpod9pmzw" satisfied condition "running" +Feb 25 10:47:24.147: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-8297 exec execpod9pmzw -- /bin/sh -x -c nslookup nodeport-service.services-8297.svc.cluster.local' +Feb 25 10:47:24.792: INFO: stderr: "+ nslookup nodeport-service.services-8297.svc.cluster.local\n" +Feb 25 10:47:24.792: INFO: stdout: "Server:\t\t10.233.0.10\nAddress:\t10.233.0.10#53\n\nnodeport-service.services-8297.svc.cluster.local\tcanonical name = externalsvc.services-8297.svc.cluster.local.\nName:\texternalsvc.services-8297.svc.cluster.local\nAddress: 10.233.15.244\n\n" +STEP: deleting ReplicationController externalsvc in namespace services-8297, will wait for the garbage collector to delete the pods 02/25/23 10:47:24.792 +Feb 25 10:47:24.926: INFO: Deleting ReplicationController externalsvc took: 76.265234ms +Feb 25 10:47:25.127: INFO: Terminating ReplicationController externalsvc pods took: 201.152572ms +Feb 25 10:47:28.223: INFO: Cleaning up the NodePort to ExternalName test service +[AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 16:37:35.682: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "svcaccounts-1380" for this suite. 12/14/22 16:37:35.689 -{"msg":"PASSED [sig-auth] ServiceAccounts should mount an API token into pods [Conformance]","completed":195,"skipped":3523,"failed":0} +Feb 25 10:47:28.293: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-8297" for this suite. 02/25/23 10:47:28.343 +[AfterEach] [sig-network] Services + test/e2e/network/service.go:762 +{"msg":"PASSED [sig-network] Services should be able to change the type from NodePort to ExternalName [Conformance]","completed":191,"skipped":3517,"failed":0} ------------------------------ -• [2.767 seconds] -[sig-auth] ServiceAccounts -test/e2e/auth/framework.go:23 - should mount an API token into pods [Conformance] - test/e2e/auth/service_accounts.go:75 +• [SLOW TEST] [14.654 seconds] +[sig-network] Services +test/e2e/network/common/framework.go:23 + should be able to change the type from NodePort to ExternalName [Conformance] + test/e2e/network/service.go:1523 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-auth] ServiceAccounts + [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:37:32.933 - Dec 14 16:37:32.933: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename svcaccounts 12/14/22 16:37:32.935 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:32.959 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:32.962 - [It] should mount an API token into pods [Conformance] - test/e2e/auth/service_accounts.go:75 - Dec 14 16:37:32.988: INFO: Waiting up to 5m0s for pod "pod-service-account-8de78ed8-a20d-451c-a251-47d57f1715dd" in namespace "svcaccounts-1380" to be "running" - Dec 14 16:37:32.998: INFO: Pod "pod-service-account-8de78ed8-a20d-451c-a251-47d57f1715dd": Phase="Pending", Reason="", readiness=false. Elapsed: 10.366424ms - Dec 14 16:37:35.009: INFO: Pod "pod-service-account-8de78ed8-a20d-451c-a251-47d57f1715dd": Phase="Running", Reason="", readiness=true. Elapsed: 2.020918856s - Dec 14 16:37:35.009: INFO: Pod "pod-service-account-8de78ed8-a20d-451c-a251-47d57f1715dd" satisfied condition "running" - STEP: reading a file in the container 12/14/22 16:37:35.009 - Dec 14 16:37:35.010: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-1380 pod-service-account-8de78ed8-a20d-451c-a251-47d57f1715dd -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token' - STEP: reading a file in the container 12/14/22 16:37:35.245 - Dec 14 16:37:35.246: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-1380 pod-service-account-8de78ed8-a20d-451c-a251-47d57f1715dd -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt' - STEP: reading a file in the container 12/14/22 16:37:35.435 - Dec 14 16:37:35.436: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-1380 pod-service-account-8de78ed8-a20d-451c-a251-47d57f1715dd -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace' - Dec 14 16:37:35.678: INFO: Got root ca configmap in namespace "svcaccounts-1380" - [AfterEach] [sig-auth] ServiceAccounts + STEP: Creating a kubernetes client 02/25/23 10:47:13.715 + Feb 25 10:47:13.716: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 10:47:13.719 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:47:13.81 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:47:13.819 + [BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 + [It] should be able to change the type from NodePort to ExternalName [Conformance] + test/e2e/network/service.go:1523 + STEP: creating a service nodeport-service with the type=NodePort in namespace services-8297 02/25/23 10:47:13.836 + STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service 02/25/23 10:47:13.888 + STEP: creating service externalsvc in namespace services-8297 02/25/23 10:47:13.888 + STEP: creating replication controller externalsvc in namespace services-8297 02/25/23 10:47:13.96 + I0225 10:47:13.981982 13 runners.go:193] Created replication controller with name: externalsvc, namespace: services-8297, replica count: 2 + I0225 10:47:17.034902 13 runners.go:193] externalsvc Pods: 2 out of 2 created, 1 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + I0225 10:47:20.035480 13 runners.go:193] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + STEP: changing the NodePort service to type=ExternalName 02/25/23 10:47:20.044 + Feb 25 10:47:20.095: INFO: Creating new exec pod + Feb 25 10:47:20.125: INFO: Waiting up to 5m0s for pod "execpod9pmzw" in namespace "services-8297" to be "running" + Feb 25 10:47:20.135: INFO: Pod "execpod9pmzw": Phase="Pending", Reason="", readiness=false. Elapsed: 9.623685ms + Feb 25 10:47:22.146: INFO: Pod "execpod9pmzw": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020591862s + Feb 25 10:47:24.146: INFO: Pod "execpod9pmzw": Phase="Running", Reason="", readiness=true. Elapsed: 4.020382222s + Feb 25 10:47:24.146: INFO: Pod "execpod9pmzw" satisfied condition "running" + Feb 25 10:47:24.147: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-8297 exec execpod9pmzw -- /bin/sh -x -c nslookup nodeport-service.services-8297.svc.cluster.local' + Feb 25 10:47:24.792: INFO: stderr: "+ nslookup nodeport-service.services-8297.svc.cluster.local\n" + Feb 25 10:47:24.792: INFO: stdout: "Server:\t\t10.233.0.10\nAddress:\t10.233.0.10#53\n\nnodeport-service.services-8297.svc.cluster.local\tcanonical name = externalsvc.services-8297.svc.cluster.local.\nName:\texternalsvc.services-8297.svc.cluster.local\nAddress: 10.233.15.244\n\n" + STEP: deleting ReplicationController externalsvc in namespace services-8297, will wait for the garbage collector to delete the pods 02/25/23 10:47:24.792 + Feb 25 10:47:24.926: INFO: Deleting ReplicationController externalsvc took: 76.265234ms + Feb 25 10:47:25.127: INFO: Terminating ReplicationController externalsvc pods took: 201.152572ms + Feb 25 10:47:28.223: INFO: Cleaning up the NodePort to ExternalName test service + [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 16:37:35.682: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "svcaccounts-1380" for this suite. 12/14/22 16:37:35.689 + Feb 25 10:47:28.293: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-8297" for this suite. 02/25/23 10:47:28.343 + [AfterEach] [sig-network] Services + test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Secrets - should patch a secret [Conformance] - test/e2e/common/node/secrets.go:153 -[BeforeEach] [sig-node] Secrets +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + removes definition from spec when one version gets changed to not be served [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:441 +[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:37:35.701 -Dec 14 16:37:35.702: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename secrets 12/14/22 16:37:35.708 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:35.733 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:35.736 -[It] should patch a secret [Conformance] - test/e2e/common/node/secrets.go:153 -STEP: creating a secret 12/14/22 16:37:35.739 -STEP: listing secrets in all namespaces to ensure that there are more than zero 12/14/22 16:37:35.744 -STEP: patching the secret 12/14/22 16:37:35.749 -STEP: deleting the secret using a LabelSelector 12/14/22 16:37:35.779 -STEP: listing secrets in all namespaces, searching for label name and value in patch 12/14/22 16:37:35.795 -[AfterEach] [sig-node] Secrets +STEP: Creating a kubernetes client 02/25/23 10:47:28.385 +Feb 25 10:47:28.386: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 10:47:28.391 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:47:28.45 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:47:28.456 +[It] removes definition from spec when one version gets changed to not be served [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:441 +STEP: set up a multi version CRD 02/25/23 10:47:28.463 +Feb 25 10:47:28.466: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: mark a version not serverd 02/25/23 10:47:49.28 +STEP: check the unserved version gets removed 02/25/23 10:47:49.363 +STEP: check the other version is not changed 02/25/23 10:47:56.759 +[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:37:35.799: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-6316" for this suite. 12/14/22 16:37:35.804 -{"msg":"PASSED [sig-node] Secrets should patch a secret [Conformance]","completed":196,"skipped":3538,"failed":0} +Feb 25 10:48:10.754: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "crd-publish-openapi-7068" for this suite. 02/25/23 10:48:10.775 +{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] removes definition from spec when one version gets changed to not be served [Conformance]","completed":192,"skipped":3537,"failed":0} ------------------------------ -• [0.113 seconds] -[sig-node] Secrets -test/e2e/common/node/framework.go:23 - should patch a secret [Conformance] - test/e2e/common/node/secrets.go:153 +• [SLOW TEST] [42.406 seconds] +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + removes definition from spec when one version gets changed to not be served [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:441 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Secrets + [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:37:35.701 - Dec 14 16:37:35.702: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename secrets 12/14/22 16:37:35.708 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:35.733 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:35.736 - [It] should patch a secret [Conformance] - test/e2e/common/node/secrets.go:153 - STEP: creating a secret 12/14/22 16:37:35.739 - STEP: listing secrets in all namespaces to ensure that there are more than zero 12/14/22 16:37:35.744 - STEP: patching the secret 12/14/22 16:37:35.749 - STEP: deleting the secret using a LabelSelector 12/14/22 16:37:35.779 - STEP: listing secrets in all namespaces, searching for label name and value in patch 12/14/22 16:37:35.795 - [AfterEach] [sig-node] Secrets + STEP: Creating a kubernetes client 02/25/23 10:47:28.385 + Feb 25 10:47:28.386: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 10:47:28.391 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:47:28.45 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:47:28.456 + [It] removes definition from spec when one version gets changed to not be served [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:441 + STEP: set up a multi version CRD 02/25/23 10:47:28.463 + Feb 25 10:47:28.466: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: mark a version not serverd 02/25/23 10:47:49.28 + STEP: check the unserved version gets removed 02/25/23 10:47:49.363 + STEP: check the other version is not changed 02/25/23 10:47:56.759 + [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:37:35.799: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "secrets-6316" for this suite. 12/14/22 16:37:35.804 + Feb 25 10:48:10.754: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "crd-publish-openapi-7068" for this suite. 02/25/23 10:48:10.775 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSS ------------------------------ -[sig-apps] Deployment - deployment should delete old replica sets [Conformance] - test/e2e/apps/deployment.go:122 -[BeforeEach] [sig-apps] Deployment +[sig-apps] Job + should manage the lifecycle of a job [Conformance] + test/e2e/apps/job.go:531 +[BeforeEach] [sig-apps] Job test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:37:35.818 -Dec 14 16:37:35.818: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename deployment 12/14/22 16:37:35.819 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:35.852 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:35.856 -[BeforeEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:91 -[It] deployment should delete old replica sets [Conformance] - test/e2e/apps/deployment.go:122 -Dec 14 16:37:35.871: INFO: Pod name cleanup-pod: Found 0 pods out of 1 -Dec 14 16:37:40.878: INFO: Pod name cleanup-pod: Found 1 pods out of 1 -STEP: ensuring each pod is running 12/14/22 16:37:40.879 -Dec 14 16:37:40.879: INFO: Creating deployment test-cleanup-deployment -STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up 12/14/22 16:37:40.894 -[AfterEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:84 -Dec 14 16:37:40.906: INFO: Deployment "test-cleanup-deployment": -&Deployment{ObjectMeta:{test-cleanup-deployment deployment-5304 eddfd7e3-382c-4ac6-9a20-bfa171cbe573 21388 1 2022-12-14 16:37:40 +0000 UTC map[name:cleanup-pod] map[] [] [] [{e2e.test Update apps/v1 2022-12-14 16:37:40 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} }]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:cleanup-pod] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003711398 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:0,Replicas:0,UpdatedReplicas:0,AvailableReplicas:0,UnavailableReplicas:0,Conditions:[]DeploymentCondition{},ReadyReplicas:0,CollisionCount:nil,},} - -Dec 14 16:37:40.910: INFO: New ReplicaSet of Deployment "test-cleanup-deployment" is nil. -[AfterEach] [sig-apps] Deployment +STEP: Creating a kubernetes client 02/25/23 10:48:10.8 +Feb 25 10:48:10.801: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename job 02/25/23 10:48:10.804 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:48:10.837 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:48:10.847 +[It] should manage the lifecycle of a job [Conformance] + test/e2e/apps/job.go:531 +STEP: Creating a suspended job 02/25/23 10:48:10.865 +STEP: Patching the Job 02/25/23 10:48:10.882 +STEP: Watching for Job to be patched 02/25/23 10:48:10.942 +Feb 25 10:48:10.948: INFO: Event ADDED observed for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk] and annotations: map[batch.kubernetes.io/job-tracking:] +Feb 25 10:48:10.949: INFO: Event MODIFIED found for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking:] +STEP: Updating the job 02/25/23 10:48:10.95 +STEP: Watching for Job to be updated 02/25/23 10:48:11.076 +Feb 25 10:48:11.080: INFO: Event MODIFIED found for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] +Feb 25 10:48:11.081: INFO: Found Job annotations: map[string]string{"batch.kubernetes.io/job-tracking":"", "updated":"true"} +STEP: Listing all Jobs with LabelSelector 02/25/23 10:48:11.081 +Feb 25 10:48:11.105: INFO: Job: e2e-nk2xk as labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] +STEP: Waiting for job to complete 02/25/23 10:48:11.105 +STEP: Delete a job collection with a labelselector 02/25/23 10:48:25.201 +STEP: Watching for Job to be deleted 02/25/23 10:48:25.275 +Feb 25 10:48:25.280: INFO: Event MODIFIED observed for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] +Feb 25 10:48:25.280: INFO: Event MODIFIED observed for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] +Feb 25 10:48:25.281: INFO: Event MODIFIED observed for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] +Feb 25 10:48:25.282: INFO: Event MODIFIED observed for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] +Feb 25 10:48:25.283: INFO: Event MODIFIED observed for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] +Feb 25 10:48:25.283: INFO: Event DELETED found for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] +STEP: Relist jobs to confirm deletion 02/25/23 10:48:25.283 +[AfterEach] [sig-apps] Job test/e2e/framework/framework.go:187 -Dec 14 16:37:40.914: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "deployment-5304" for this suite. 12/14/22 16:37:40.919 -{"msg":"PASSED [sig-apps] Deployment deployment should delete old replica sets [Conformance]","completed":197,"skipped":3594,"failed":0} +Feb 25 10:48:25.297: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "job-3774" for this suite. 02/25/23 10:48:25.353 +{"msg":"PASSED [sig-apps] Job should manage the lifecycle of a job [Conformance]","completed":193,"skipped":3540,"failed":0} ------------------------------ -• [SLOW TEST] [5.108 seconds] -[sig-apps] Deployment +• [SLOW TEST] [14.665 seconds] +[sig-apps] Job test/e2e/apps/framework.go:23 - deployment should delete old replica sets [Conformance] - test/e2e/apps/deployment.go:122 + should manage the lifecycle of a job [Conformance] + test/e2e/apps/job.go:531 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Deployment + [BeforeEach] [sig-apps] Job test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:37:35.818 - Dec 14 16:37:35.818: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename deployment 12/14/22 16:37:35.819 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:35.852 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:35.856 - [BeforeEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:91 - [It] deployment should delete old replica sets [Conformance] - test/e2e/apps/deployment.go:122 - Dec 14 16:37:35.871: INFO: Pod name cleanup-pod: Found 0 pods out of 1 - Dec 14 16:37:40.878: INFO: Pod name cleanup-pod: Found 1 pods out of 1 - STEP: ensuring each pod is running 12/14/22 16:37:40.879 - Dec 14 16:37:40.879: INFO: Creating deployment test-cleanup-deployment - STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up 12/14/22 16:37:40.894 - [AfterEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:84 - Dec 14 16:37:40.906: INFO: Deployment "test-cleanup-deployment": - &Deployment{ObjectMeta:{test-cleanup-deployment deployment-5304 eddfd7e3-382c-4ac6-9a20-bfa171cbe573 21388 1 2022-12-14 16:37:40 +0000 UTC map[name:cleanup-pod] map[] [] [] [{e2e.test Update apps/v1 2022-12-14 16:37:40 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} }]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:cleanup-pod] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003711398 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:0,Replicas:0,UpdatedReplicas:0,AvailableReplicas:0,UnavailableReplicas:0,Conditions:[]DeploymentCondition{},ReadyReplicas:0,CollisionCount:nil,},} - - Dec 14 16:37:40.910: INFO: New ReplicaSet of Deployment "test-cleanup-deployment" is nil. - [AfterEach] [sig-apps] Deployment + STEP: Creating a kubernetes client 02/25/23 10:48:10.8 + Feb 25 10:48:10.801: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename job 02/25/23 10:48:10.804 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:48:10.837 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:48:10.847 + [It] should manage the lifecycle of a job [Conformance] + test/e2e/apps/job.go:531 + STEP: Creating a suspended job 02/25/23 10:48:10.865 + STEP: Patching the Job 02/25/23 10:48:10.882 + STEP: Watching for Job to be patched 02/25/23 10:48:10.942 + Feb 25 10:48:10.948: INFO: Event ADDED observed for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk] and annotations: map[batch.kubernetes.io/job-tracking:] + Feb 25 10:48:10.949: INFO: Event MODIFIED found for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking:] + STEP: Updating the job 02/25/23 10:48:10.95 + STEP: Watching for Job to be updated 02/25/23 10:48:11.076 + Feb 25 10:48:11.080: INFO: Event MODIFIED found for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] + Feb 25 10:48:11.081: INFO: Found Job annotations: map[string]string{"batch.kubernetes.io/job-tracking":"", "updated":"true"} + STEP: Listing all Jobs with LabelSelector 02/25/23 10:48:11.081 + Feb 25 10:48:11.105: INFO: Job: e2e-nk2xk as labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] + STEP: Waiting for job to complete 02/25/23 10:48:11.105 + STEP: Delete a job collection with a labelselector 02/25/23 10:48:25.201 + STEP: Watching for Job to be deleted 02/25/23 10:48:25.275 + Feb 25 10:48:25.280: INFO: Event MODIFIED observed for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] + Feb 25 10:48:25.280: INFO: Event MODIFIED observed for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] + Feb 25 10:48:25.281: INFO: Event MODIFIED observed for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] + Feb 25 10:48:25.282: INFO: Event MODIFIED observed for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] + Feb 25 10:48:25.283: INFO: Event MODIFIED observed for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] + Feb 25 10:48:25.283: INFO: Event DELETED found for Job e2e-nk2xk in namespace job-3774 with labels: map[e2e-job-label:e2e-nk2xk e2e-nk2xk:patched] and annotations: map[batch.kubernetes.io/job-tracking: updated:true] + STEP: Relist jobs to confirm deletion 02/25/23 10:48:25.283 + [AfterEach] [sig-apps] Job test/e2e/framework/framework.go:187 - Dec 14 16:37:40.914: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "deployment-5304" for this suite. 12/14/22 16:37:40.919 + Feb 25 10:48:25.297: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "job-3774" for this suite. 02/25/23 10:48:25.353 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-scheduling] SchedulerPredicates [Serial] - validates that NodeSelector is respected if matching [Conformance] - test/e2e/scheduling/predicates.go:461 -[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] +[sig-node] Downward API + should provide pod UID as env vars [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:266 +[BeforeEach] [sig-node] Downward API test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:37:40.929 -Dec 14 16:37:40.930: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename sched-pred 12/14/22 16:37:40.932 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:40.959 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:40.962 -[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:92 -Dec 14 16:37:40.966: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready -Dec 14 16:37:40.976: INFO: Waiting for terminating namespaces to be deleted... -Dec 14 16:37:40.985: INFO: -Logging pods the apiserver thinks is on node kiedai4iakae-1 before test -Dec 14 16:37:41.000: INFO: kube-flannel-ds-d6vqv from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.000: INFO: Container kube-flannel ready: true, restart count 0 -Dec 14 16:37:41.000: INFO: coredns-565d847f94-6mrkc from kube-system started at 2022-12-14 16:08:53 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.000: INFO: Container coredns ready: true, restart count 0 -Dec 14 16:37:41.000: INFO: kube-addon-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.000: INFO: Container kube-addon-manager ready: true, restart count 1 -Dec 14 16:37:41.000: INFO: kube-apiserver-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.001: INFO: Container kube-apiserver ready: true, restart count 1 -Dec 14 16:37:41.001: INFO: kube-controller-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:29:20 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.001: INFO: Container kube-controller-manager ready: true, restart count 2 -Dec 14 16:37:41.001: INFO: kube-proxy-fs5fm from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.001: INFO: Container kube-proxy ready: true, restart count 0 -Dec 14 16:37:41.001: INFO: kube-scheduler-kiedai4iakae-1 from kube-system started at 2022-12-14 15:25:59 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.001: INFO: Container kube-scheduler ready: true, restart count 2 -Dec 14 16:37:41.001: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-z7zks from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 16:37:41.001: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 16:37:41.001: INFO: Container systemd-logs ready: true, restart count 0 -Dec 14 16:37:41.001: INFO: -Logging pods the apiserver thinks is on node kiedai4iakae-2 before test -Dec 14 16:37:41.014: INFO: kube-flannel-ds-nvdqm from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.014: INFO: Container kube-flannel ready: true, restart count 0 -Dec 14 16:37:41.014: INFO: coredns-565d847f94-xhv9r from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.014: INFO: Container coredns ready: true, restart count 0 -Dec 14 16:37:41.014: INFO: kube-addon-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.014: INFO: Container kube-addon-manager ready: true, restart count 1 -Dec 14 16:37:41.014: INFO: kube-apiserver-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.014: INFO: Container kube-apiserver ready: true, restart count 1 -Dec 14 16:37:41.014: INFO: kube-controller-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.014: INFO: Container kube-controller-manager ready: true, restart count 1 -Dec 14 16:37:41.014: INFO: kube-proxy-dqjw5 from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.014: INFO: Container kube-proxy ready: true, restart count 0 -Dec 14 16:37:41.014: INFO: kube-scheduler-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.014: INFO: Container kube-scheduler ready: true, restart count 2 -Dec 14 16:37:41.014: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-cgz78 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 16:37:41.014: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 16:37:41.014: INFO: Container systemd-logs ready: true, restart count 0 -Dec 14 16:37:41.014: INFO: -Logging pods the apiserver thinks is on node kiedai4iakae-3 before test -Dec 14 16:37:41.028: INFO: test-cleanup-controller-zdt25 from deployment-5304 started at 2022-12-14 16:37:36 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.028: INFO: Container httpd ready: true, restart count 0 -Dec 14 16:37:41.028: INFO: kube-flannel-ds-4ws4h from kube-flannel started at 2022-12-14 16:08:54 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.028: INFO: Container kube-flannel ready: true, restart count 0 -Dec 14 16:37:41.028: INFO: kube-proxy-8kpmb from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.028: INFO: Container kube-proxy ready: true, restart count 0 -Dec 14 16:37:41.028: INFO: sonobuoy from sonobuoy started at 2022-12-14 15:35:59 +0000 UTC (1 container statuses recorded) -Dec 14 16:37:41.028: INFO: Container kube-sonobuoy ready: true, restart count 0 -Dec 14 16:37:41.028: INFO: sonobuoy-e2e-job-177ae5e9cdd34bac from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 16:37:41.028: INFO: Container e2e ready: true, restart count 0 -Dec 14 16:37:41.028: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 16:37:41.028: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-2fkm5 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) -Dec 14 16:37:41.028: INFO: Container sonobuoy-worker ready: true, restart count 0 -Dec 14 16:37:41.028: INFO: Container systemd-logs ready: true, restart count 0 -[It] validates that NodeSelector is respected if matching [Conformance] - test/e2e/scheduling/predicates.go:461 -STEP: Trying to launch a pod without a label to get a node which can launch it. 12/14/22 16:37:41.028 -Dec 14 16:37:41.039: INFO: Waiting up to 1m0s for pod "without-label" in namespace "sched-pred-3441" to be "running" -Dec 14 16:37:41.043: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 4.425926ms -Dec 14 16:37:43.052: INFO: Pod "without-label": Phase="Running", Reason="", readiness=true. Elapsed: 2.012774763s -Dec 14 16:37:43.052: INFO: Pod "without-label" satisfied condition "running" -STEP: Explicitly delete pod here to free the resource it takes. 12/14/22 16:37:43.058 -STEP: Trying to apply a random label on the found node. 12/14/22 16:37:43.078 -STEP: verifying the node has the label kubernetes.io/e2e-0c7e20c4-77f0-484c-b055-3227d6a6e559 42 12/14/22 16:37:43.103 -STEP: Trying to relaunch the pod, now with labels. 12/14/22 16:37:43.109 -Dec 14 16:37:43.118: INFO: Waiting up to 5m0s for pod "with-labels" in namespace "sched-pred-3441" to be "not pending" -Dec 14 16:37:43.124: INFO: Pod "with-labels": Phase="Pending", Reason="", readiness=false. Elapsed: 5.521963ms -Dec 14 16:37:45.136: INFO: Pod "with-labels": Phase="Running", Reason="", readiness=true. Elapsed: 2.018059884s -Dec 14 16:37:45.137: INFO: Pod "with-labels" satisfied condition "not pending" -STEP: removing the label kubernetes.io/e2e-0c7e20c4-77f0-484c-b055-3227d6a6e559 off the node kiedai4iakae-3 12/14/22 16:37:45.143 -STEP: verifying the node doesn't have the label kubernetes.io/e2e-0c7e20c4-77f0-484c-b055-3227d6a6e559 12/14/22 16:37:45.167 -[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] +STEP: Creating a kubernetes client 02/25/23 10:48:25.482 +Feb 25 10:48:25.483: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 10:48:25.491 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:48:25.573 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:48:25.579 +[It] should provide pod UID as env vars [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:266 +STEP: Creating a pod to test downward api env vars 02/25/23 10:48:25.585 +Feb 25 10:48:25.620: INFO: Waiting up to 5m0s for pod "downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79" in namespace "downward-api-9717" to be "Succeeded or Failed" +Feb 25 10:48:25.654: INFO: Pod "downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79": Phase="Pending", Reason="", readiness=false. Elapsed: 33.125593ms +Feb 25 10:48:27.666: INFO: Pod "downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79": Phase="Pending", Reason="", readiness=false. Elapsed: 2.045685627s +Feb 25 10:48:29.665: INFO: Pod "downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79": Phase="Pending", Reason="", readiness=false. Elapsed: 4.04451974s +Feb 25 10:48:31.663: INFO: Pod "downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.042157134s +STEP: Saw pod success 02/25/23 10:48:31.663 +Feb 25 10:48:31.663: INFO: Pod "downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79" satisfied condition "Succeeded or Failed" +Feb 25 10:48:31.675: INFO: Trying to get logs from node diek3dua9gay-3 pod downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79 container dapi-container: +STEP: delete the pod 02/25/23 10:48:31.902 +Feb 25 10:48:31.929: INFO: Waiting for pod downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79 to disappear +Feb 25 10:48:31.937: INFO: Pod downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79 no longer exists +[AfterEach] [sig-node] Downward API test/e2e/framework/framework.go:187 -Dec 14 16:37:45.174: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "sched-pred-3441" for this suite. 12/14/22 16:37:45.179 -[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:83 -{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching [Conformance]","completed":198,"skipped":3597,"failed":0} +Feb 25 10:48:31.937: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-9717" for this suite. 02/25/23 10:48:31.945 +{"msg":"PASSED [sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance]","completed":194,"skipped":3563,"failed":0} ------------------------------ -• [4.258 seconds] -[sig-scheduling] SchedulerPredicates [Serial] -test/e2e/scheduling/framework.go:40 - validates that NodeSelector is respected if matching [Conformance] - test/e2e/scheduling/predicates.go:461 +• [SLOW TEST] [6.481 seconds] +[sig-node] Downward API +test/e2e/common/node/framework.go:23 + should provide pod UID as env vars [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:266 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + [BeforeEach] [sig-node] Downward API test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:37:40.929 - Dec 14 16:37:40.930: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename sched-pred 12/14/22 16:37:40.932 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:40.959 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:40.962 - [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:92 - Dec 14 16:37:40.966: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready - Dec 14 16:37:40.976: INFO: Waiting for terminating namespaces to be deleted... - Dec 14 16:37:40.985: INFO: - Logging pods the apiserver thinks is on node kiedai4iakae-1 before test - Dec 14 16:37:41.000: INFO: kube-flannel-ds-d6vqv from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.000: INFO: Container kube-flannel ready: true, restart count 0 - Dec 14 16:37:41.000: INFO: coredns-565d847f94-6mrkc from kube-system started at 2022-12-14 16:08:53 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.000: INFO: Container coredns ready: true, restart count 0 - Dec 14 16:37:41.000: INFO: kube-addon-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.000: INFO: Container kube-addon-manager ready: true, restart count 1 - Dec 14 16:37:41.000: INFO: kube-apiserver-kiedai4iakae-1 from kube-system started at 2022-12-14 15:34:43 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.001: INFO: Container kube-apiserver ready: true, restart count 1 - Dec 14 16:37:41.001: INFO: kube-controller-manager-kiedai4iakae-1 from kube-system started at 2022-12-14 15:29:20 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.001: INFO: Container kube-controller-manager ready: true, restart count 2 - Dec 14 16:37:41.001: INFO: kube-proxy-fs5fm from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.001: INFO: Container kube-proxy ready: true, restart count 0 - Dec 14 16:37:41.001: INFO: kube-scheduler-kiedai4iakae-1 from kube-system started at 2022-12-14 15:25:59 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.001: INFO: Container kube-scheduler ready: true, restart count 2 - Dec 14 16:37:41.001: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-z7zks from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 16:37:41.001: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 16:37:41.001: INFO: Container systemd-logs ready: true, restart count 0 - Dec 14 16:37:41.001: INFO: - Logging pods the apiserver thinks is on node kiedai4iakae-2 before test - Dec 14 16:37:41.014: INFO: kube-flannel-ds-nvdqm from kube-flannel started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.014: INFO: Container kube-flannel ready: true, restart count 0 - Dec 14 16:37:41.014: INFO: coredns-565d847f94-xhv9r from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.014: INFO: Container coredns ready: true, restart count 0 - Dec 14 16:37:41.014: INFO: kube-addon-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.014: INFO: Container kube-addon-manager ready: true, restart count 1 - Dec 14 16:37:41.014: INFO: kube-apiserver-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.014: INFO: Container kube-apiserver ready: true, restart count 1 - Dec 14 16:37:41.014: INFO: kube-controller-manager-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.014: INFO: Container kube-controller-manager ready: true, restart count 1 - Dec 14 16:37:41.014: INFO: kube-proxy-dqjw5 from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.014: INFO: Container kube-proxy ready: true, restart count 0 - Dec 14 16:37:41.014: INFO: kube-scheduler-kiedai4iakae-2 from kube-system started at 2022-12-14 15:34:57 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.014: INFO: Container kube-scheduler ready: true, restart count 2 - Dec 14 16:37:41.014: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-cgz78 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 16:37:41.014: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 16:37:41.014: INFO: Container systemd-logs ready: true, restart count 0 - Dec 14 16:37:41.014: INFO: - Logging pods the apiserver thinks is on node kiedai4iakae-3 before test - Dec 14 16:37:41.028: INFO: test-cleanup-controller-zdt25 from deployment-5304 started at 2022-12-14 16:37:36 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.028: INFO: Container httpd ready: true, restart count 0 - Dec 14 16:37:41.028: INFO: kube-flannel-ds-4ws4h from kube-flannel started at 2022-12-14 16:08:54 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.028: INFO: Container kube-flannel ready: true, restart count 0 - Dec 14 16:37:41.028: INFO: kube-proxy-8kpmb from kube-system started at 2022-12-14 15:35:23 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.028: INFO: Container kube-proxy ready: true, restart count 0 - Dec 14 16:37:41.028: INFO: sonobuoy from sonobuoy started at 2022-12-14 15:35:59 +0000 UTC (1 container statuses recorded) - Dec 14 16:37:41.028: INFO: Container kube-sonobuoy ready: true, restart count 0 - Dec 14 16:37:41.028: INFO: sonobuoy-e2e-job-177ae5e9cdd34bac from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 16:37:41.028: INFO: Container e2e ready: true, restart count 0 - Dec 14 16:37:41.028: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 16:37:41.028: INFO: sonobuoy-systemd-logs-daemon-set-44851d2dd865472c-2fkm5 from sonobuoy started at 2022-12-14 15:36:09 +0000 UTC (2 container statuses recorded) - Dec 14 16:37:41.028: INFO: Container sonobuoy-worker ready: true, restart count 0 - Dec 14 16:37:41.028: INFO: Container systemd-logs ready: true, restart count 0 - [It] validates that NodeSelector is respected if matching [Conformance] - test/e2e/scheduling/predicates.go:461 - STEP: Trying to launch a pod without a label to get a node which can launch it. 12/14/22 16:37:41.028 - Dec 14 16:37:41.039: INFO: Waiting up to 1m0s for pod "without-label" in namespace "sched-pred-3441" to be "running" - Dec 14 16:37:41.043: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 4.425926ms - Dec 14 16:37:43.052: INFO: Pod "without-label": Phase="Running", Reason="", readiness=true. Elapsed: 2.012774763s - Dec 14 16:37:43.052: INFO: Pod "without-label" satisfied condition "running" - STEP: Explicitly delete pod here to free the resource it takes. 12/14/22 16:37:43.058 - STEP: Trying to apply a random label on the found node. 12/14/22 16:37:43.078 - STEP: verifying the node has the label kubernetes.io/e2e-0c7e20c4-77f0-484c-b055-3227d6a6e559 42 12/14/22 16:37:43.103 - STEP: Trying to relaunch the pod, now with labels. 12/14/22 16:37:43.109 - Dec 14 16:37:43.118: INFO: Waiting up to 5m0s for pod "with-labels" in namespace "sched-pred-3441" to be "not pending" - Dec 14 16:37:43.124: INFO: Pod "with-labels": Phase="Pending", Reason="", readiness=false. Elapsed: 5.521963ms - Dec 14 16:37:45.136: INFO: Pod "with-labels": Phase="Running", Reason="", readiness=true. Elapsed: 2.018059884s - Dec 14 16:37:45.137: INFO: Pod "with-labels" satisfied condition "not pending" - STEP: removing the label kubernetes.io/e2e-0c7e20c4-77f0-484c-b055-3227d6a6e559 off the node kiedai4iakae-3 12/14/22 16:37:45.143 - STEP: verifying the node doesn't have the label kubernetes.io/e2e-0c7e20c4-77f0-484c-b055-3227d6a6e559 12/14/22 16:37:45.167 - [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + STEP: Creating a kubernetes client 02/25/23 10:48:25.482 + Feb 25 10:48:25.483: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 10:48:25.491 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:48:25.573 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:48:25.579 + [It] should provide pod UID as env vars [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:266 + STEP: Creating a pod to test downward api env vars 02/25/23 10:48:25.585 + Feb 25 10:48:25.620: INFO: Waiting up to 5m0s for pod "downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79" in namespace "downward-api-9717" to be "Succeeded or Failed" + Feb 25 10:48:25.654: INFO: Pod "downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79": Phase="Pending", Reason="", readiness=false. Elapsed: 33.125593ms + Feb 25 10:48:27.666: INFO: Pod "downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79": Phase="Pending", Reason="", readiness=false. Elapsed: 2.045685627s + Feb 25 10:48:29.665: INFO: Pod "downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79": Phase="Pending", Reason="", readiness=false. Elapsed: 4.04451974s + Feb 25 10:48:31.663: INFO: Pod "downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.042157134s + STEP: Saw pod success 02/25/23 10:48:31.663 + Feb 25 10:48:31.663: INFO: Pod "downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79" satisfied condition "Succeeded or Failed" + Feb 25 10:48:31.675: INFO: Trying to get logs from node diek3dua9gay-3 pod downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79 container dapi-container: + STEP: delete the pod 02/25/23 10:48:31.902 + Feb 25 10:48:31.929: INFO: Waiting for pod downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79 to disappear + Feb 25 10:48:31.937: INFO: Pod downward-api-b4d149d2-2007-45e6-9cb9-a1d2cf522a79 no longer exists + [AfterEach] [sig-node] Downward API test/e2e/framework/framework.go:187 - Dec 14 16:37:45.174: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "sched-pred-3441" for this suite. 12/14/22 16:37:45.179 - [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] - test/e2e/scheduling/predicates.go:83 + Feb 25 10:48:31.937: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-9717" for this suite. 02/25/23 10:48:31.945 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSS ------------------------------ [sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance] test/e2e/common/node/configmap.go:137 [BeforeEach] [sig-node] ConfigMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:37:45.19 -Dec 14 16:37:45.191: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename configmap 12/14/22 16:37:45.192 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:45.214 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:45.219 +STEP: Creating a kubernetes client 02/25/23 10:48:31.967 +Feb 25 10:48:31.967: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename configmap 02/25/23 10:48:31.971 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:48:32 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:48:32.006 [It] should fail to create ConfigMap with empty key [Conformance] test/e2e/common/node/configmap.go:137 -STEP: Creating configMap that has name configmap-test-emptyKey-17171d6b-0fe4-4d50-96ed-63e63d4b74af 12/14/22 16:37:45.223 +STEP: Creating configMap that has name configmap-test-emptyKey-f7cb404f-4a25-48fd-8877-81463c6de625 02/25/23 10:48:32.011 [AfterEach] [sig-node] ConfigMap test/e2e/framework/framework.go:187 -Dec 14 16:37:45.226: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-4208" for this suite. 12/14/22 16:37:45.232 -{"msg":"PASSED [sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance]","completed":199,"skipped":3634,"failed":0} +Feb 25 10:48:32.014: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-2459" for this suite. 02/25/23 10:48:32.028 +{"msg":"PASSED [sig-node] ConfigMap should fail to create ConfigMap with empty key [Conformance]","completed":195,"skipped":3579,"failed":0} ------------------------------ -• [0.052 seconds] +• [0.078 seconds] [sig-node] ConfigMap test/e2e/common/node/framework.go:23 should fail to create ConfigMap with empty key [Conformance] @@ -18431,3930 +17851,5204 @@ test/e2e/common/node/framework.go:23 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-node] ConfigMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:37:45.19 - Dec 14 16:37:45.191: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename configmap 12/14/22 16:37:45.192 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:45.214 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:45.219 + STEP: Creating a kubernetes client 02/25/23 10:48:31.967 + Feb 25 10:48:31.967: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename configmap 02/25/23 10:48:31.971 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:48:32 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:48:32.006 [It] should fail to create ConfigMap with empty key [Conformance] test/e2e/common/node/configmap.go:137 - STEP: Creating configMap that has name configmap-test-emptyKey-17171d6b-0fe4-4d50-96ed-63e63d4b74af 12/14/22 16:37:45.223 + STEP: Creating configMap that has name configmap-test-emptyKey-f7cb404f-4a25-48fd-8877-81463c6de625 02/25/23 10:48:32.011 [AfterEach] [sig-node] ConfigMap test/e2e/framework/framework.go:187 - Dec 14 16:37:45.226: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "configmap-4208" for this suite. 12/14/22 16:37:45.232 + Feb 25 10:48:32.014: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "configmap-2459" for this suite. 02/25/23 10:48:32.028 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSS ------------------------------ -[sig-node] Variable Expansion - should allow substituting values in a container's command [NodeConformance] [Conformance] - test/e2e/common/node/expansion.go:72 -[BeforeEach] [sig-node] Variable Expansion +[sig-node] ConfigMap + should be consumable via the environment [NodeConformance] [Conformance] + test/e2e/common/node/configmap.go:92 +[BeforeEach] [sig-node] ConfigMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:37:45.25 -Dec 14 16:37:45.251: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename var-expansion 12/14/22 16:37:45.252 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:45.273 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:45.277 -[It] should allow substituting values in a container's command [NodeConformance] [Conformance] - test/e2e/common/node/expansion.go:72 -STEP: Creating a pod to test substitution in container's command 12/14/22 16:37:45.284 -Dec 14 16:37:45.315: INFO: Waiting up to 5m0s for pod "var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2" in namespace "var-expansion-9027" to be "Succeeded or Failed" -Dec 14 16:37:45.324: INFO: Pod "var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2": Phase="Pending", Reason="", readiness=false. Elapsed: 9.25743ms -Dec 14 16:37:47.330: INFO: Pod "var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015327215s -Dec 14 16:37:49.331: INFO: Pod "var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2": Phase="Pending", Reason="", readiness=false. Elapsed: 4.01661766s -Dec 14 16:37:51.330: INFO: Pod "var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.015189912s -STEP: Saw pod success 12/14/22 16:37:51.33 -Dec 14 16:37:51.331: INFO: Pod "var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2" satisfied condition "Succeeded or Failed" -Dec 14 16:37:51.335: INFO: Trying to get logs from node kiedai4iakae-3 pod var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2 container dapi-container: -STEP: delete the pod 12/14/22 16:37:51.348 -Dec 14 16:37:51.379: INFO: Waiting for pod var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2 to disappear -Dec 14 16:37:51.392: INFO: Pod var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2 no longer exists -[AfterEach] [sig-node] Variable Expansion +STEP: Creating a kubernetes client 02/25/23 10:48:32.067 +Feb 25 10:48:32.067: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename configmap 02/25/23 10:48:32.069 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:48:32.098 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:48:32.104 +[It] should be consumable via the environment [NodeConformance] [Conformance] + test/e2e/common/node/configmap.go:92 +STEP: Creating configMap configmap-6140/configmap-test-e468384a-791d-4b95-8f6f-5e8313f40f6b 02/25/23 10:48:32.11 +STEP: Creating a pod to test consume configMaps 02/25/23 10:48:32.119 +Feb 25 10:48:32.131: INFO: Waiting up to 5m0s for pod "pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a" in namespace "configmap-6140" to be "Succeeded or Failed" +Feb 25 10:48:32.140: INFO: Pod "pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a": Phase="Pending", Reason="", readiness=false. Elapsed: 8.843646ms +Feb 25 10:48:34.151: INFO: Pod "pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020101673s +Feb 25 10:48:36.148: INFO: Pod "pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.017119847s +Feb 25 10:48:38.150: INFO: Pod "pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.018525985s +STEP: Saw pod success 02/25/23 10:48:38.15 +Feb 25 10:48:38.150: INFO: Pod "pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a" satisfied condition "Succeeded or Failed" +Feb 25 10:48:38.157: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a container env-test: +STEP: delete the pod 02/25/23 10:48:38.188 +Feb 25 10:48:38.236: INFO: Waiting for pod pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a to disappear +Feb 25 10:48:38.248: INFO: Pod pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a no longer exists +[AfterEach] [sig-node] ConfigMap test/e2e/framework/framework.go:187 -Dec 14 16:37:51.392: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "var-expansion-9027" for this suite. 12/14/22 16:37:51.401 -{"msg":"PASSED [sig-node] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance]","completed":200,"skipped":3664,"failed":0} +Feb 25 10:48:38.249: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-6140" for this suite. 02/25/23 10:48:38.262 +{"msg":"PASSED [sig-node] ConfigMap should be consumable via the environment [NodeConformance] [Conformance]","completed":196,"skipped":3588,"failed":0} ------------------------------ -• [SLOW TEST] [6.159 seconds] -[sig-node] Variable Expansion +• [SLOW TEST] [6.238 seconds] +[sig-node] ConfigMap test/e2e/common/node/framework.go:23 - should allow substituting values in a container's command [NodeConformance] [Conformance] - test/e2e/common/node/expansion.go:72 + should be consumable via the environment [NodeConformance] [Conformance] + test/e2e/common/node/configmap.go:92 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Variable Expansion + [BeforeEach] [sig-node] ConfigMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:37:45.25 - Dec 14 16:37:45.251: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename var-expansion 12/14/22 16:37:45.252 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:45.273 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:45.277 - [It] should allow substituting values in a container's command [NodeConformance] [Conformance] - test/e2e/common/node/expansion.go:72 - STEP: Creating a pod to test substitution in container's command 12/14/22 16:37:45.284 - Dec 14 16:37:45.315: INFO: Waiting up to 5m0s for pod "var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2" in namespace "var-expansion-9027" to be "Succeeded or Failed" - Dec 14 16:37:45.324: INFO: Pod "var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2": Phase="Pending", Reason="", readiness=false. Elapsed: 9.25743ms - Dec 14 16:37:47.330: INFO: Pod "var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015327215s - Dec 14 16:37:49.331: INFO: Pod "var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2": Phase="Pending", Reason="", readiness=false. Elapsed: 4.01661766s - Dec 14 16:37:51.330: INFO: Pod "var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.015189912s - STEP: Saw pod success 12/14/22 16:37:51.33 - Dec 14 16:37:51.331: INFO: Pod "var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2" satisfied condition "Succeeded or Failed" - Dec 14 16:37:51.335: INFO: Trying to get logs from node kiedai4iakae-3 pod var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2 container dapi-container: - STEP: delete the pod 12/14/22 16:37:51.348 - Dec 14 16:37:51.379: INFO: Waiting for pod var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2 to disappear - Dec 14 16:37:51.392: INFO: Pod var-expansion-6a212601-18bb-4706-9738-a28d8f571ed2 no longer exists - [AfterEach] [sig-node] Variable Expansion + STEP: Creating a kubernetes client 02/25/23 10:48:32.067 + Feb 25 10:48:32.067: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename configmap 02/25/23 10:48:32.069 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:48:32.098 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:48:32.104 + [It] should be consumable via the environment [NodeConformance] [Conformance] + test/e2e/common/node/configmap.go:92 + STEP: Creating configMap configmap-6140/configmap-test-e468384a-791d-4b95-8f6f-5e8313f40f6b 02/25/23 10:48:32.11 + STEP: Creating a pod to test consume configMaps 02/25/23 10:48:32.119 + Feb 25 10:48:32.131: INFO: Waiting up to 5m0s for pod "pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a" in namespace "configmap-6140" to be "Succeeded or Failed" + Feb 25 10:48:32.140: INFO: Pod "pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a": Phase="Pending", Reason="", readiness=false. Elapsed: 8.843646ms + Feb 25 10:48:34.151: INFO: Pod "pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.020101673s + Feb 25 10:48:36.148: INFO: Pod "pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.017119847s + Feb 25 10:48:38.150: INFO: Pod "pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.018525985s + STEP: Saw pod success 02/25/23 10:48:38.15 + Feb 25 10:48:38.150: INFO: Pod "pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a" satisfied condition "Succeeded or Failed" + Feb 25 10:48:38.157: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a container env-test: + STEP: delete the pod 02/25/23 10:48:38.188 + Feb 25 10:48:38.236: INFO: Waiting for pod pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a to disappear + Feb 25 10:48:38.248: INFO: Pod pod-configmaps-3f71a525-6e1c-4853-91da-629639c4659a no longer exists + [AfterEach] [sig-node] ConfigMap test/e2e/framework/framework.go:187 - Dec 14 16:37:51.392: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "var-expansion-9027" for this suite. 12/14/22 16:37:51.401 + Feb 25 10:48:38.249: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "configmap-6140" for this suite. 02/25/23 10:48:38.262 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:166 -[BeforeEach] [sig-storage] EmptyDir volumes +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + patching/updating a validating webhook should work [Conformance] + test/e2e/apimachinery/webhook.go:412 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:37:51.41 -Dec 14 16:37:51.410: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir 12/14/22 16:37:51.412 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:51.498 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:51.502 -[It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:166 -STEP: Creating a pod to test emptydir 0644 on node default medium 12/14/22 16:37:51.504 -Dec 14 16:37:51.515: INFO: Waiting up to 5m0s for pod "pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a" in namespace "emptydir-6595" to be "Succeeded or Failed" -Dec 14 16:37:51.530: INFO: Pod "pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a": Phase="Pending", Reason="", readiness=false. Elapsed: 14.570687ms -Dec 14 16:37:53.540: INFO: Pod "pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024636172s -Dec 14 16:37:55.539: INFO: Pod "pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.02346082s -STEP: Saw pod success 12/14/22 16:37:55.539 -Dec 14 16:37:55.539: INFO: Pod "pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a" satisfied condition "Succeeded or Failed" -Dec 14 16:37:55.545: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a container test-container: -STEP: delete the pod 12/14/22 16:37:55.555 -Dec 14 16:37:55.575: INFO: Waiting for pod pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a to disappear -Dec 14 16:37:55.580: INFO: Pod pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +STEP: Creating a kubernetes client 02/25/23 10:48:38.308 +Feb 25 10:48:38.309: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 10:48:38.314 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:48:38.365 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:48:38.372 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 +STEP: Setting up server cert 02/25/23 10:48:38.41 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:48:40.382 +STEP: Deploying the webhook pod 02/25/23 10:48:40.398 +STEP: Wait for the deployment to be ready 02/25/23 10:48:40.421 +Feb 25 10:48:40.453: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set +Feb 25 10:48:42.480: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 48, 40, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 48, 40, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 48, 40, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 48, 40, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} +STEP: Deploying the webhook service 02/25/23 10:48:44.489 +STEP: Verifying the service has paired with the endpoint 02/25/23 10:48:44.514 +Feb 25 10:48:45.514: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] patching/updating a validating webhook should work [Conformance] + test/e2e/apimachinery/webhook.go:412 +STEP: Creating a validating webhook configuration 02/25/23 10:48:45.546 +Feb 25 10:48:45.600: INFO: Waiting for webhook configuration to be ready... +STEP: Creating a configMap that does not comply to the validation webhook rules 02/25/23 10:48:45.726 +STEP: Updating a validating webhook configuration's rules to not include the create operation 02/25/23 10:48:45.794 +STEP: Creating a configMap that does not comply to the validation webhook rules 02/25/23 10:48:45.819 +STEP: Patching a validating webhook configuration's rules to include the create operation 02/25/23 10:48:45.874 +STEP: Creating a configMap that does not comply to the validation webhook rules 02/25/23 10:48:45.903 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:37:55.580: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-6595" for this suite. 12/14/22 16:37:55.589 -{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","completed":201,"skipped":3677,"failed":0} +Feb 25 10:48:45.927: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-8266" for this suite. 02/25/23 10:48:45.939 +STEP: Destroying namespace "webhook-8266-markers" for this suite. 02/25/23 10:48:45.968 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a validating webhook should work [Conformance]","completed":197,"skipped":3620,"failed":0} ------------------------------ -• [4.190 seconds] -[sig-storage] EmptyDir volumes +• [SLOW TEST] [7.935 seconds] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + patching/updating a validating webhook should work [Conformance] + test/e2e/apimachinery/webhook.go:412 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:48:38.308 + Feb 25 10:48:38.309: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 10:48:38.314 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:48:38.365 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:48:38.372 + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 + STEP: Setting up server cert 02/25/23 10:48:38.41 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 10:48:40.382 + STEP: Deploying the webhook pod 02/25/23 10:48:40.398 + STEP: Wait for the deployment to be ready 02/25/23 10:48:40.421 + Feb 25 10:48:40.453: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set + Feb 25 10:48:42.480: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 10, 48, 40, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 48, 40, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 48, 40, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 48, 40, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} + STEP: Deploying the webhook service 02/25/23 10:48:44.489 + STEP: Verifying the service has paired with the endpoint 02/25/23 10:48:44.514 + Feb 25 10:48:45.514: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] patching/updating a validating webhook should work [Conformance] + test/e2e/apimachinery/webhook.go:412 + STEP: Creating a validating webhook configuration 02/25/23 10:48:45.546 + Feb 25 10:48:45.600: INFO: Waiting for webhook configuration to be ready... + STEP: Creating a configMap that does not comply to the validation webhook rules 02/25/23 10:48:45.726 + STEP: Updating a validating webhook configuration's rules to not include the create operation 02/25/23 10:48:45.794 + STEP: Creating a configMap that does not comply to the validation webhook rules 02/25/23 10:48:45.819 + STEP: Patching a validating webhook configuration's rules to include the create operation 02/25/23 10:48:45.874 + STEP: Creating a configMap that does not comply to the validation webhook rules 02/25/23 10:48:45.903 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/framework/framework.go:187 + Feb 25 10:48:45.927: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-8266" for this suite. 02/25/23 10:48:45.939 + STEP: Destroying namespace "webhook-8266-markers" for this suite. 02/25/23 10:48:45.968 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-network] DNS + should provide DNS for the cluster [Conformance] + test/e2e/network/dns.go:50 +[BeforeEach] [sig-network] DNS + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:48:46.276 +Feb 25 10:48:46.276: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename dns 02/25/23 10:48:46.279 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:48:46.342 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:48:46.353 +[It] should provide DNS for the cluster [Conformance] + test/e2e/network/dns.go:50 +STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;sleep 1; done + 02/25/23 10:48:46.363 +STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;sleep 1; done + 02/25/23 10:48:46.365 +STEP: creating a pod to probe DNS 02/25/23 10:48:46.365 +STEP: submitting the pod to kubernetes 02/25/23 10:48:46.365 +Feb 25 10:48:46.391: INFO: Waiting up to 15m0s for pod "dns-test-f8fe1a8b-0946-4465-a9ba-ba92a39790a0" in namespace "dns-4837" to be "running" +Feb 25 10:48:46.399: INFO: Pod "dns-test-f8fe1a8b-0946-4465-a9ba-ba92a39790a0": Phase="Pending", Reason="", readiness=false. Elapsed: 8.237898ms +Feb 25 10:48:48.415: INFO: Pod "dns-test-f8fe1a8b-0946-4465-a9ba-ba92a39790a0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023449554s +Feb 25 10:48:50.409: INFO: Pod "dns-test-f8fe1a8b-0946-4465-a9ba-ba92a39790a0": Phase="Pending", Reason="", readiness=false. Elapsed: 4.017538237s +Feb 25 10:48:52.419: INFO: Pod "dns-test-f8fe1a8b-0946-4465-a9ba-ba92a39790a0": Phase="Running", Reason="", readiness=true. Elapsed: 6.02729408s +Feb 25 10:48:52.419: INFO: Pod "dns-test-f8fe1a8b-0946-4465-a9ba-ba92a39790a0" satisfied condition "running" +STEP: retrieving the pod 02/25/23 10:48:52.419 +STEP: looking for the results for each expected name from probers 02/25/23 10:48:52.443 +Feb 25 10:48:52.558: INFO: DNS probes using dns-4837/dns-test-f8fe1a8b-0946-4465-a9ba-ba92a39790a0 succeeded + +STEP: deleting the pod 02/25/23 10:48:52.558 +[AfterEach] [sig-network] DNS + test/e2e/framework/framework.go:187 +Feb 25 10:48:52.615: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "dns-4837" for this suite. 02/25/23 10:48:52.628 +{"msg":"PASSED [sig-network] DNS should provide DNS for the cluster [Conformance]","completed":198,"skipped":3666,"failed":0} +------------------------------ +• [SLOW TEST] [6.379 seconds] +[sig-network] DNS +test/e2e/network/common/framework.go:23 + should provide DNS for the cluster [Conformance] + test/e2e/network/dns.go:50 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-network] DNS + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:48:46.276 + Feb 25 10:48:46.276: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename dns 02/25/23 10:48:46.279 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:48:46.342 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:48:46.353 + [It] should provide DNS for the cluster [Conformance] + test/e2e/network/dns.go:50 + STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@kubernetes.default.svc.cluster.local;sleep 1; done + 02/25/23 10:48:46.363 + STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@kubernetes.default.svc.cluster.local;check="$$(dig +tcp +noall +answer +search kubernetes.default.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@kubernetes.default.svc.cluster.local;sleep 1; done + 02/25/23 10:48:46.365 + STEP: creating a pod to probe DNS 02/25/23 10:48:46.365 + STEP: submitting the pod to kubernetes 02/25/23 10:48:46.365 + Feb 25 10:48:46.391: INFO: Waiting up to 15m0s for pod "dns-test-f8fe1a8b-0946-4465-a9ba-ba92a39790a0" in namespace "dns-4837" to be "running" + Feb 25 10:48:46.399: INFO: Pod "dns-test-f8fe1a8b-0946-4465-a9ba-ba92a39790a0": Phase="Pending", Reason="", readiness=false. Elapsed: 8.237898ms + Feb 25 10:48:48.415: INFO: Pod "dns-test-f8fe1a8b-0946-4465-a9ba-ba92a39790a0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023449554s + Feb 25 10:48:50.409: INFO: Pod "dns-test-f8fe1a8b-0946-4465-a9ba-ba92a39790a0": Phase="Pending", Reason="", readiness=false. Elapsed: 4.017538237s + Feb 25 10:48:52.419: INFO: Pod "dns-test-f8fe1a8b-0946-4465-a9ba-ba92a39790a0": Phase="Running", Reason="", readiness=true. Elapsed: 6.02729408s + Feb 25 10:48:52.419: INFO: Pod "dns-test-f8fe1a8b-0946-4465-a9ba-ba92a39790a0" satisfied condition "running" + STEP: retrieving the pod 02/25/23 10:48:52.419 + STEP: looking for the results for each expected name from probers 02/25/23 10:48:52.443 + Feb 25 10:48:52.558: INFO: DNS probes using dns-4837/dns-test-f8fe1a8b-0946-4465-a9ba-ba92a39790a0 succeeded + + STEP: deleting the pod 02/25/23 10:48:52.558 + [AfterEach] [sig-network] DNS + test/e2e/framework/framework.go:187 + Feb 25 10:48:52.615: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "dns-4837" for this suite. 02/25/23 10:48:52.628 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSSSSS +------------------------------ +[sig-storage] ConfigMap + should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:422 +[BeforeEach] [sig-storage] ConfigMap + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:48:52.656 +Feb 25 10:48:52.656: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename configmap 02/25/23 10:48:52.677 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:48:52.752 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:48:52.758 +[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:422 +STEP: Creating configMap with name configmap-test-volume-02c6a6ce-ae47-4d77-8fbd-ee35424f53e6 02/25/23 10:48:52.763 +STEP: Creating a pod to test consume configMaps 02/25/23 10:48:52.781 +Feb 25 10:48:52.803: INFO: Waiting up to 5m0s for pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f" in namespace "configmap-5755" to be "Succeeded or Failed" +Feb 25 10:48:52.825: INFO: Pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f": Phase="Pending", Reason="", readiness=false. Elapsed: 21.726208ms +Feb 25 10:48:54.838: INFO: Pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.035571461s +Feb 25 10:48:56.857: INFO: Pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.054098968s +Feb 25 10:48:58.840: INFO: Pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f": Phase="Pending", Reason="", readiness=false. Elapsed: 6.037064264s +Feb 25 10:49:00.834: INFO: Pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f": Phase="Pending", Reason="", readiness=false. Elapsed: 8.031508585s +Feb 25 10:49:02.836: INFO: Pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.033245564s +STEP: Saw pod success 02/25/23 10:49:02.836 +Feb 25 10:49:02.837: INFO: Pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f" satisfied condition "Succeeded or Failed" +Feb 25 10:49:02.845: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f container configmap-volume-test: +STEP: delete the pod 02/25/23 10:49:02.86 +Feb 25 10:49:02.898: INFO: Waiting for pod pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f to disappear +Feb 25 10:49:02.909: INFO: Pod pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f no longer exists +[AfterEach] [sig-storage] ConfigMap + test/e2e/framework/framework.go:187 +Feb 25 10:49:02.910: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-5755" for this suite. 02/25/23 10:49:02.955 +{"msg":"PASSED [sig-storage] ConfigMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","completed":199,"skipped":3684,"failed":0} +------------------------------ +• [SLOW TEST] [10.331 seconds] +[sig-storage] ConfigMap test/e2e/common/storage/framework.go:23 - should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:166 + should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:422 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir volumes + [BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:37:51.41 - Dec 14 16:37:51.410: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir 12/14/22 16:37:51.412 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:51.498 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:51.502 - [It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:166 - STEP: Creating a pod to test emptydir 0644 on node default medium 12/14/22 16:37:51.504 - Dec 14 16:37:51.515: INFO: Waiting up to 5m0s for pod "pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a" in namespace "emptydir-6595" to be "Succeeded or Failed" - Dec 14 16:37:51.530: INFO: Pod "pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a": Phase="Pending", Reason="", readiness=false. Elapsed: 14.570687ms - Dec 14 16:37:53.540: INFO: Pod "pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024636172s - Dec 14 16:37:55.539: INFO: Pod "pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.02346082s - STEP: Saw pod success 12/14/22 16:37:55.539 - Dec 14 16:37:55.539: INFO: Pod "pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a" satisfied condition "Succeeded or Failed" - Dec 14 16:37:55.545: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a container test-container: - STEP: delete the pod 12/14/22 16:37:55.555 - Dec 14 16:37:55.575: INFO: Waiting for pod pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a to disappear - Dec 14 16:37:55.580: INFO: Pod pod-598ba1a1-ecb2-4423-8fb1-0b7d3a587e3a no longer exists - [AfterEach] [sig-storage] EmptyDir volumes + STEP: Creating a kubernetes client 02/25/23 10:48:52.656 + Feb 25 10:48:52.656: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename configmap 02/25/23 10:48:52.677 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:48:52.752 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:48:52.758 + [It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:422 + STEP: Creating configMap with name configmap-test-volume-02c6a6ce-ae47-4d77-8fbd-ee35424f53e6 02/25/23 10:48:52.763 + STEP: Creating a pod to test consume configMaps 02/25/23 10:48:52.781 + Feb 25 10:48:52.803: INFO: Waiting up to 5m0s for pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f" in namespace "configmap-5755" to be "Succeeded or Failed" + Feb 25 10:48:52.825: INFO: Pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f": Phase="Pending", Reason="", readiness=false. Elapsed: 21.726208ms + Feb 25 10:48:54.838: INFO: Pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.035571461s + Feb 25 10:48:56.857: INFO: Pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.054098968s + Feb 25 10:48:58.840: INFO: Pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f": Phase="Pending", Reason="", readiness=false. Elapsed: 6.037064264s + Feb 25 10:49:00.834: INFO: Pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f": Phase="Pending", Reason="", readiness=false. Elapsed: 8.031508585s + Feb 25 10:49:02.836: INFO: Pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.033245564s + STEP: Saw pod success 02/25/23 10:49:02.836 + Feb 25 10:49:02.837: INFO: Pod "pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f" satisfied condition "Succeeded or Failed" + Feb 25 10:49:02.845: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f container configmap-volume-test: + STEP: delete the pod 02/25/23 10:49:02.86 + Feb 25 10:49:02.898: INFO: Waiting for pod pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f to disappear + Feb 25 10:49:02.909: INFO: Pod pod-configmaps-693026e5-a0c0-4697-8b3b-a1c9bf78d65f no longer exists + [AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 - Dec 14 16:37:55.580: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-6595" for this suite. 12/14/22 16:37:55.589 + Feb 25 10:49:02.910: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "configmap-5755" for this suite. 02/25/23 10:49:02.955 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - works for CRD without validation schema [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:152 -[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +[sig-scheduling] SchedulerPredicates [Serial] + validates that NodeSelector is respected if matching [Conformance] + test/e2e/scheduling/predicates.go:461 +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:37:55.611 -Dec 14 16:37:55.611: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 16:37:55.612 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:55.643 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:55.645 -[It] works for CRD without validation schema [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:152 -Dec 14 16:37:55.650: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: kubectl validation (kubectl create and apply) allows request with any unknown properties 12/14/22 16:37:59.187 -Dec 14 16:37:59.188: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-6041 --namespace=crd-publish-openapi-6041 create -f -' -Dec 14 16:38:00.588: INFO: stderr: "" -Dec 14 16:38:00.588: INFO: stdout: "e2e-test-crd-publish-openapi-3270-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" -Dec 14 16:38:00.589: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-6041 --namespace=crd-publish-openapi-6041 delete e2e-test-crd-publish-openapi-3270-crds test-cr' -Dec 14 16:38:00.805: INFO: stderr: "" -Dec 14 16:38:00.805: INFO: stdout: "e2e-test-crd-publish-openapi-3270-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" -Dec 14 16:38:00.807: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-6041 --namespace=crd-publish-openapi-6041 apply -f -' -Dec 14 16:38:01.872: INFO: stderr: "" -Dec 14 16:38:01.872: INFO: stdout: "e2e-test-crd-publish-openapi-3270-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" -Dec 14 16:38:01.872: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-6041 --namespace=crd-publish-openapi-6041 delete e2e-test-crd-publish-openapi-3270-crds test-cr' -Dec 14 16:38:02.032: INFO: stderr: "" -Dec 14 16:38:02.032: INFO: stdout: "e2e-test-crd-publish-openapi-3270-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" -STEP: kubectl explain works to explain CR without validation schema 12/14/22 16:38:02.032 -Dec 14 16:38:02.033: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-6041 explain e2e-test-crd-publish-openapi-3270-crds' -Dec 14 16:38:02.471: INFO: stderr: "" -Dec 14 16:38:02.471: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-3270-crd\nVERSION: crd-publish-openapi-test-empty.example.com/v1\n\nDESCRIPTION:\n \n" -[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 10:49:02.987 +Feb 25 10:49:02.988: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename sched-pred 02/25/23 10:49:03.006 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:03.058 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:03.067 +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:92 +Feb 25 10:49:03.078: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready +Feb 25 10:49:03.117: INFO: Waiting for terminating namespaces to be deleted... +Feb 25 10:49:03.125: INFO: +Logging pods the apiserver thinks is on node diek3dua9gay-1 before test +Feb 25 10:49:03.153: INFO: echo-other-node-cd69fcf6b-hsnk9 from cilium-test started at 2023-02-25 09:38:06 +0000 UTC (2 container statuses recorded) +Feb 25 10:49:03.153: INFO: Container dns-test-server ready: true, restart count 0 +Feb 25 10:49:03.153: INFO: Container echo-other-node ready: true, restart count 0 +Feb 25 10:49:03.153: INFO: cilium-node-init-s84z5 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.153: INFO: Container node-init ready: true, restart count 0 +Feb 25 10:49:03.153: INFO: cilium-r4lc4 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.153: INFO: Container cilium-agent ready: true, restart count 0 +Feb 25 10:49:03.153: INFO: coredns-565d847f94-qqvf6 from kube-system started at 2023-02-25 08:49:06 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.153: INFO: Container coredns ready: true, restart count 0 +Feb 25 10:49:03.153: INFO: kube-addon-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.153: INFO: Container kube-addon-manager ready: true, restart count 0 +Feb 25 10:49:03.153: INFO: kube-apiserver-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:09 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.153: INFO: Container kube-apiserver ready: true, restart count 0 +Feb 25 10:49:03.153: INFO: kube-controller-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:49:55 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.154: INFO: Container kube-controller-manager ready: true, restart count 0 +Feb 25 10:49:03.154: INFO: kube-proxy-cm4qg from kube-system started at 2023-02-25 08:46:38 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.154: INFO: Container kube-proxy ready: true, restart count 0 +Feb 25 10:49:03.154: INFO: kube-scheduler-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:27 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.154: INFO: Container kube-scheduler ready: true, restart count 0 +Feb 25 10:49:03.154: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-wznrx from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 10:49:03.154: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 10:49:03.154: INFO: Container systemd-logs ready: true, restart count 0 +Feb 25 10:49:03.154: INFO: +Logging pods the apiserver thinks is on node diek3dua9gay-2 before test +Feb 25 10:49:03.217: INFO: cilium-9tkfg from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.217: INFO: Container cilium-agent ready: true, restart count 0 +Feb 25 10:49:03.217: INFO: cilium-node-init-8b7n2 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.217: INFO: Container node-init ready: true, restart count 0 +Feb 25 10:49:03.217: INFO: cilium-operator-59d7d77cdc-lppbd from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.217: INFO: Container cilium-operator ready: true, restart count 0 +Feb 25 10:49:03.217: INFO: coredns-565d847f94-5nh2n from kube-system started at 2023-02-25 08:48:51 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.217: INFO: Container coredns ready: true, restart count 0 +Feb 25 10:49:03.217: INFO: kube-addon-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.217: INFO: Container kube-addon-manager ready: true, restart count 0 +Feb 25 10:49:03.217: INFO: kube-apiserver-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.217: INFO: Container kube-apiserver ready: true, restart count 0 +Feb 25 10:49:03.217: INFO: kube-controller-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.217: INFO: Container kube-controller-manager ready: true, restart count 0 +Feb 25 10:49:03.217: INFO: kube-proxy-d66m7 from kube-system started at 2023-02-25 08:46:57 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.217: INFO: Container kube-proxy ready: true, restart count 0 +Feb 25 10:49:03.217: INFO: kube-scheduler-diek3dua9gay-2 from kube-system started at 2023-02-25 08:50:15 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.217: INFO: Container kube-scheduler ready: true, restart count 0 +Feb 25 10:49:03.217: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-mml8w from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 10:49:03.217: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 10:49:03.217: INFO: Container systemd-logs ready: true, restart count 0 +Feb 25 10:49:03.217: INFO: +Logging pods the apiserver thinks is on node diek3dua9gay-3 before test +Feb 25 10:49:03.265: INFO: client-7b78db77d5-27xbt from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.266: INFO: Container client ready: true, restart count 0 +Feb 25 10:49:03.266: INFO: client2-78f748dd67-skwz2 from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.266: INFO: Container client2 ready: true, restart count 0 +Feb 25 10:49:03.266: INFO: echo-same-node-85bc9b6b56-wqmfq from cilium-test started at 2023-02-25 09:57:46 +0000 UTC (2 container statuses recorded) +Feb 25 10:49:03.266: INFO: Container dns-test-server ready: true, restart count 0 +Feb 25 10:49:03.266: INFO: Container echo-same-node ready: true, restart count 0 +Feb 25 10:49:03.266: INFO: cilium-2cncz from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.266: INFO: Container cilium-agent ready: true, restart count 0 +Feb 25 10:49:03.266: INFO: cilium-node-init-979mb from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.266: INFO: Container node-init ready: true, restart count 0 +Feb 25 10:49:03.266: INFO: kube-proxy-22s9q from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.266: INFO: Container kube-proxy ready: true, restart count 0 +Feb 25 10:49:03.266: INFO: sonobuoy from sonobuoy started at 2023-02-25 09:45:14 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:03.266: INFO: Container kube-sonobuoy ready: true, restart count 0 +Feb 25 10:49:03.266: INFO: sonobuoy-e2e-job-0b3c30975b5c4880 from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 10:49:03.266: INFO: Container e2e ready: true, restart count 0 +Feb 25 10:49:03.266: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 10:49:03.267: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-9pdjb from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 10:49:03.267: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 10:49:03.268: INFO: Container systemd-logs ready: true, restart count 0 +[It] validates that NodeSelector is respected if matching [Conformance] + test/e2e/scheduling/predicates.go:461 +STEP: Trying to launch a pod without a label to get a node which can launch it. 02/25/23 10:49:03.268 +Feb 25 10:49:03.300: INFO: Waiting up to 1m0s for pod "without-label" in namespace "sched-pred-545" to be "running" +Feb 25 10:49:03.316: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 16.214929ms +Feb 25 10:49:05.329: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029228299s +Feb 25 10:49:07.343: INFO: Pod "without-label": Phase="Running", Reason="", readiness=true. Elapsed: 4.042703893s +Feb 25 10:49:07.343: INFO: Pod "without-label" satisfied condition "running" +STEP: Explicitly delete pod here to free the resource it takes. 02/25/23 10:49:07.379 +STEP: Trying to apply a random label on the found node. 02/25/23 10:49:07.452 +STEP: verifying the node has the label kubernetes.io/e2e-68f3282e-f680-4dc0-9998-3b44436d2708 42 02/25/23 10:49:07.675 +STEP: Trying to relaunch the pod, now with labels. 02/25/23 10:49:07.714 +Feb 25 10:49:07.765: INFO: Waiting up to 5m0s for pod "with-labels" in namespace "sched-pred-545" to be "not pending" +Feb 25 10:49:07.860: INFO: Pod "with-labels": Phase="Pending", Reason="", readiness=false. Elapsed: 94.52985ms +Feb 25 10:49:09.871: INFO: Pod "with-labels": Phase="Pending", Reason="", readiness=false. Elapsed: 2.105652662s +Feb 25 10:49:11.868: INFO: Pod "with-labels": Phase="Running", Reason="", readiness=true. Elapsed: 4.102743908s +Feb 25 10:49:11.868: INFO: Pod "with-labels" satisfied condition "not pending" +STEP: removing the label kubernetes.io/e2e-68f3282e-f680-4dc0-9998-3b44436d2708 off the node diek3dua9gay-3 02/25/23 10:49:11.876 +STEP: verifying the node doesn't have the label kubernetes.io/e2e-68f3282e-f680-4dc0-9998-3b44436d2708 02/25/23 10:49:11.913 +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:187 -Dec 14 16:38:05.614: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-publish-openapi-6041" for this suite. 12/14/22 16:38:05.627 -{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD without validation schema [Conformance]","completed":202,"skipped":3787,"failed":0} +Feb 25 10:49:11.928: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "sched-pred-545" for this suite. 02/25/23 10:49:11.942 +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:83 +{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if matching [Conformance]","completed":200,"skipped":3684,"failed":0} ------------------------------ -• [SLOW TEST] [10.024 seconds] -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - works for CRD without validation schema [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:152 +• [SLOW TEST] [8.978 seconds] +[sig-scheduling] SchedulerPredicates [Serial] +test/e2e/scheduling/framework.go:40 + validates that NodeSelector is respected if matching [Conformance] + test/e2e/scheduling/predicates.go:461 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:37:55.611 - Dec 14 16:37:55.611: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 16:37:55.612 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:37:55.643 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:37:55.645 - [It] works for CRD without validation schema [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:152 - Dec 14 16:37:55.650: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: kubectl validation (kubectl create and apply) allows request with any unknown properties 12/14/22 16:37:59.187 - Dec 14 16:37:59.188: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-6041 --namespace=crd-publish-openapi-6041 create -f -' - Dec 14 16:38:00.588: INFO: stderr: "" - Dec 14 16:38:00.588: INFO: stdout: "e2e-test-crd-publish-openapi-3270-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" - Dec 14 16:38:00.589: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-6041 --namespace=crd-publish-openapi-6041 delete e2e-test-crd-publish-openapi-3270-crds test-cr' - Dec 14 16:38:00.805: INFO: stderr: "" - Dec 14 16:38:00.805: INFO: stdout: "e2e-test-crd-publish-openapi-3270-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" - Dec 14 16:38:00.807: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-6041 --namespace=crd-publish-openapi-6041 apply -f -' - Dec 14 16:38:01.872: INFO: stderr: "" - Dec 14 16:38:01.872: INFO: stdout: "e2e-test-crd-publish-openapi-3270-crd.crd-publish-openapi-test-empty.example.com/test-cr created\n" - Dec 14 16:38:01.872: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-6041 --namespace=crd-publish-openapi-6041 delete e2e-test-crd-publish-openapi-3270-crds test-cr' - Dec 14 16:38:02.032: INFO: stderr: "" - Dec 14 16:38:02.032: INFO: stdout: "e2e-test-crd-publish-openapi-3270-crd.crd-publish-openapi-test-empty.example.com \"test-cr\" deleted\n" - STEP: kubectl explain works to explain CR without validation schema 12/14/22 16:38:02.032 - Dec 14 16:38:02.033: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=crd-publish-openapi-6041 explain e2e-test-crd-publish-openapi-3270-crds' - Dec 14 16:38:02.471: INFO: stderr: "" - Dec 14 16:38:02.471: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-3270-crd\nVERSION: crd-publish-openapi-test-empty.example.com/v1\n\nDESCRIPTION:\n \n" - [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 10:49:02.987 + Feb 25 10:49:02.988: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename sched-pred 02/25/23 10:49:03.006 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:03.058 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:03.067 + [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:92 + Feb 25 10:49:03.078: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready + Feb 25 10:49:03.117: INFO: Waiting for terminating namespaces to be deleted... + Feb 25 10:49:03.125: INFO: + Logging pods the apiserver thinks is on node diek3dua9gay-1 before test + Feb 25 10:49:03.153: INFO: echo-other-node-cd69fcf6b-hsnk9 from cilium-test started at 2023-02-25 09:38:06 +0000 UTC (2 container statuses recorded) + Feb 25 10:49:03.153: INFO: Container dns-test-server ready: true, restart count 0 + Feb 25 10:49:03.153: INFO: Container echo-other-node ready: true, restart count 0 + Feb 25 10:49:03.153: INFO: cilium-node-init-s84z5 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.153: INFO: Container node-init ready: true, restart count 0 + Feb 25 10:49:03.153: INFO: cilium-r4lc4 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.153: INFO: Container cilium-agent ready: true, restart count 0 + Feb 25 10:49:03.153: INFO: coredns-565d847f94-qqvf6 from kube-system started at 2023-02-25 08:49:06 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.153: INFO: Container coredns ready: true, restart count 0 + Feb 25 10:49:03.153: INFO: kube-addon-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.153: INFO: Container kube-addon-manager ready: true, restart count 0 + Feb 25 10:49:03.153: INFO: kube-apiserver-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:09 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.153: INFO: Container kube-apiserver ready: true, restart count 0 + Feb 25 10:49:03.153: INFO: kube-controller-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:49:55 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.154: INFO: Container kube-controller-manager ready: true, restart count 0 + Feb 25 10:49:03.154: INFO: kube-proxy-cm4qg from kube-system started at 2023-02-25 08:46:38 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.154: INFO: Container kube-proxy ready: true, restart count 0 + Feb 25 10:49:03.154: INFO: kube-scheduler-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:27 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.154: INFO: Container kube-scheduler ready: true, restart count 0 + Feb 25 10:49:03.154: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-wznrx from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 10:49:03.154: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 10:49:03.154: INFO: Container systemd-logs ready: true, restart count 0 + Feb 25 10:49:03.154: INFO: + Logging pods the apiserver thinks is on node diek3dua9gay-2 before test + Feb 25 10:49:03.217: INFO: cilium-9tkfg from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.217: INFO: Container cilium-agent ready: true, restart count 0 + Feb 25 10:49:03.217: INFO: cilium-node-init-8b7n2 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.217: INFO: Container node-init ready: true, restart count 0 + Feb 25 10:49:03.217: INFO: cilium-operator-59d7d77cdc-lppbd from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.217: INFO: Container cilium-operator ready: true, restart count 0 + Feb 25 10:49:03.217: INFO: coredns-565d847f94-5nh2n from kube-system started at 2023-02-25 08:48:51 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.217: INFO: Container coredns ready: true, restart count 0 + Feb 25 10:49:03.217: INFO: kube-addon-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.217: INFO: Container kube-addon-manager ready: true, restart count 0 + Feb 25 10:49:03.217: INFO: kube-apiserver-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.217: INFO: Container kube-apiserver ready: true, restart count 0 + Feb 25 10:49:03.217: INFO: kube-controller-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.217: INFO: Container kube-controller-manager ready: true, restart count 0 + Feb 25 10:49:03.217: INFO: kube-proxy-d66m7 from kube-system started at 2023-02-25 08:46:57 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.217: INFO: Container kube-proxy ready: true, restart count 0 + Feb 25 10:49:03.217: INFO: kube-scheduler-diek3dua9gay-2 from kube-system started at 2023-02-25 08:50:15 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.217: INFO: Container kube-scheduler ready: true, restart count 0 + Feb 25 10:49:03.217: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-mml8w from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 10:49:03.217: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 10:49:03.217: INFO: Container systemd-logs ready: true, restart count 0 + Feb 25 10:49:03.217: INFO: + Logging pods the apiserver thinks is on node diek3dua9gay-3 before test + Feb 25 10:49:03.265: INFO: client-7b78db77d5-27xbt from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.266: INFO: Container client ready: true, restart count 0 + Feb 25 10:49:03.266: INFO: client2-78f748dd67-skwz2 from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.266: INFO: Container client2 ready: true, restart count 0 + Feb 25 10:49:03.266: INFO: echo-same-node-85bc9b6b56-wqmfq from cilium-test started at 2023-02-25 09:57:46 +0000 UTC (2 container statuses recorded) + Feb 25 10:49:03.266: INFO: Container dns-test-server ready: true, restart count 0 + Feb 25 10:49:03.266: INFO: Container echo-same-node ready: true, restart count 0 + Feb 25 10:49:03.266: INFO: cilium-2cncz from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.266: INFO: Container cilium-agent ready: true, restart count 0 + Feb 25 10:49:03.266: INFO: cilium-node-init-979mb from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.266: INFO: Container node-init ready: true, restart count 0 + Feb 25 10:49:03.266: INFO: kube-proxy-22s9q from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.266: INFO: Container kube-proxy ready: true, restart count 0 + Feb 25 10:49:03.266: INFO: sonobuoy from sonobuoy started at 2023-02-25 09:45:14 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:03.266: INFO: Container kube-sonobuoy ready: true, restart count 0 + Feb 25 10:49:03.266: INFO: sonobuoy-e2e-job-0b3c30975b5c4880 from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 10:49:03.266: INFO: Container e2e ready: true, restart count 0 + Feb 25 10:49:03.266: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 10:49:03.267: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-9pdjb from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 10:49:03.267: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 10:49:03.268: INFO: Container systemd-logs ready: true, restart count 0 + [It] validates that NodeSelector is respected if matching [Conformance] + test/e2e/scheduling/predicates.go:461 + STEP: Trying to launch a pod without a label to get a node which can launch it. 02/25/23 10:49:03.268 + Feb 25 10:49:03.300: INFO: Waiting up to 1m0s for pod "without-label" in namespace "sched-pred-545" to be "running" + Feb 25 10:49:03.316: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 16.214929ms + Feb 25 10:49:05.329: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029228299s + Feb 25 10:49:07.343: INFO: Pod "without-label": Phase="Running", Reason="", readiness=true. Elapsed: 4.042703893s + Feb 25 10:49:07.343: INFO: Pod "without-label" satisfied condition "running" + STEP: Explicitly delete pod here to free the resource it takes. 02/25/23 10:49:07.379 + STEP: Trying to apply a random label on the found node. 02/25/23 10:49:07.452 + STEP: verifying the node has the label kubernetes.io/e2e-68f3282e-f680-4dc0-9998-3b44436d2708 42 02/25/23 10:49:07.675 + STEP: Trying to relaunch the pod, now with labels. 02/25/23 10:49:07.714 + Feb 25 10:49:07.765: INFO: Waiting up to 5m0s for pod "with-labels" in namespace "sched-pred-545" to be "not pending" + Feb 25 10:49:07.860: INFO: Pod "with-labels": Phase="Pending", Reason="", readiness=false. Elapsed: 94.52985ms + Feb 25 10:49:09.871: INFO: Pod "with-labels": Phase="Pending", Reason="", readiness=false. Elapsed: 2.105652662s + Feb 25 10:49:11.868: INFO: Pod "with-labels": Phase="Running", Reason="", readiness=true. Elapsed: 4.102743908s + Feb 25 10:49:11.868: INFO: Pod "with-labels" satisfied condition "not pending" + STEP: removing the label kubernetes.io/e2e-68f3282e-f680-4dc0-9998-3b44436d2708 off the node diek3dua9gay-3 02/25/23 10:49:11.876 + STEP: verifying the node doesn't have the label kubernetes.io/e2e-68f3282e-f680-4dc0-9998-3b44436d2708 02/25/23 10:49:11.913 + [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:187 - Dec 14 16:38:05.614: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "crd-publish-openapi-6041" for this suite. 12/14/22 16:38:05.627 + Feb 25 10:49:11.928: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "sched-pred-545" for this suite. 02/25/23 10:49:11.942 + [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:83 << End Captured GinkgoWriter Output ------------------------------ -SSSSS +SSS ------------------------------ -[sig-node] Kubelet when scheduling a busybox command that always fails in a pod - should have an terminated reason [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:110 -[BeforeEach] [sig-node] Kubelet +[sig-instrumentation] Events + should delete a collection of events [Conformance] + test/e2e/instrumentation/core_events.go:175 +[BeforeEach] [sig-instrumentation] Events test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:38:05.636 -Dec 14 16:38:05.636: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubelet-test 12/14/22 16:38:05.638 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:38:05.667 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:38:05.669 -[BeforeEach] [sig-node] Kubelet - test/e2e/common/node/kubelet.go:41 -[BeforeEach] when scheduling a busybox command that always fails in a pod - test/e2e/common/node/kubelet.go:85 -[It] should have an terminated reason [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:110 -[AfterEach] [sig-node] Kubelet +STEP: Creating a kubernetes client 02/25/23 10:49:11.968 +Feb 25 10:49:11.969: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename events 02/25/23 10:49:11.98 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:12.021 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:12.034 +[It] should delete a collection of events [Conformance] + test/e2e/instrumentation/core_events.go:175 +STEP: Create set of events 02/25/23 10:49:12.044 +Feb 25 10:49:12.057: INFO: created test-event-1 +Feb 25 10:49:12.066: INFO: created test-event-2 +Feb 25 10:49:12.088: INFO: created test-event-3 +STEP: get a list of Events with a label in the current namespace 02/25/23 10:49:12.088 +STEP: delete collection of events 02/25/23 10:49:12.103 +Feb 25 10:49:12.103: INFO: requesting DeleteCollection of events +STEP: check that the list of events matches the requested quantity 02/25/23 10:49:12.158 +Feb 25 10:49:12.159: INFO: requesting list of events to confirm quantity +[AfterEach] [sig-instrumentation] Events test/e2e/framework/framework.go:187 -Dec 14 16:38:09.702: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubelet-test-894" for this suite. 12/14/22 16:38:09.714 -{"msg":"PASSED [sig-node] Kubelet when scheduling a busybox command that always fails in a pod should have an terminated reason [NodeConformance] [Conformance]","completed":203,"skipped":3792,"failed":0} +Feb 25 10:49:12.181: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "events-5475" for this suite. 02/25/23 10:49:12.199 +{"msg":"PASSED [sig-instrumentation] Events should delete a collection of events [Conformance]","completed":201,"skipped":3687,"failed":0} ------------------------------ -• [4.087 seconds] -[sig-node] Kubelet -test/e2e/common/node/framework.go:23 - when scheduling a busybox command that always fails in a pod - test/e2e/common/node/kubelet.go:82 - should have an terminated reason [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:110 +• [0.245 seconds] +[sig-instrumentation] Events +test/e2e/instrumentation/common/framework.go:23 + should delete a collection of events [Conformance] + test/e2e/instrumentation/core_events.go:175 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Kubelet + [BeforeEach] [sig-instrumentation] Events test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:38:05.636 - Dec 14 16:38:05.636: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubelet-test 12/14/22 16:38:05.638 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:38:05.667 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:38:05.669 - [BeforeEach] [sig-node] Kubelet - test/e2e/common/node/kubelet.go:41 - [BeforeEach] when scheduling a busybox command that always fails in a pod - test/e2e/common/node/kubelet.go:85 - [It] should have an terminated reason [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:110 - [AfterEach] [sig-node] Kubelet + STEP: Creating a kubernetes client 02/25/23 10:49:11.968 + Feb 25 10:49:11.969: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename events 02/25/23 10:49:11.98 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:12.021 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:12.034 + [It] should delete a collection of events [Conformance] + test/e2e/instrumentation/core_events.go:175 + STEP: Create set of events 02/25/23 10:49:12.044 + Feb 25 10:49:12.057: INFO: created test-event-1 + Feb 25 10:49:12.066: INFO: created test-event-2 + Feb 25 10:49:12.088: INFO: created test-event-3 + STEP: get a list of Events with a label in the current namespace 02/25/23 10:49:12.088 + STEP: delete collection of events 02/25/23 10:49:12.103 + Feb 25 10:49:12.103: INFO: requesting DeleteCollection of events + STEP: check that the list of events matches the requested quantity 02/25/23 10:49:12.158 + Feb 25 10:49:12.159: INFO: requesting list of events to confirm quantity + [AfterEach] [sig-instrumentation] Events test/e2e/framework/framework.go:187 - Dec 14 16:38:09.702: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubelet-test-894" for this suite. 12/14/22 16:38:09.714 + Feb 25 10:49:12.181: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "events-5475" for this suite. 02/25/23 10:49:12.199 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Pods - should run through the lifecycle of Pods and PodStatus [Conformance] - test/e2e/common/node/pods.go:895 -[BeforeEach] [sig-node] Pods +[sig-node] Variable Expansion + should allow substituting values in a container's command [NodeConformance] [Conformance] + test/e2e/common/node/expansion.go:72 +[BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:38:09.725 -Dec 14 16:38:09.725: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename pods 12/14/22 16:38:09.73 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:38:09.755 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:38:09.76 -[BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 -[It] should run through the lifecycle of Pods and PodStatus [Conformance] - test/e2e/common/node/pods.go:895 -STEP: creating a Pod with a static label 12/14/22 16:38:09.772 -STEP: watching for Pod to be ready 12/14/22 16:38:09.787 -Dec 14 16:38:09.789: INFO: observed Pod pod-test in namespace pods-2359 in phase Pending with labels: map[test-pod-static:true] & conditions [] -Dec 14 16:38:09.800: INFO: observed Pod pod-test in namespace pods-2359 in phase Pending with labels: map[test-pod-static:true] & conditions [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:09 +0000 UTC }] -Dec 14 16:38:10.224: INFO: observed Pod pod-test in namespace pods-2359 in phase Pending with labels: map[test-pod-static:true] & conditions [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:09 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:09 +0000 UTC ContainersNotReady containers with unready status: [pod-test]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:09 +0000 UTC ContainersNotReady containers with unready status: [pod-test]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:09 +0000 UTC }] -Dec 14 16:38:11.500: INFO: Found Pod pod-test in namespace pods-2359 in phase Running with labels: map[test-pod-static:true] & conditions [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:09 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:11 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:11 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:09 +0000 UTC }] -STEP: patching the Pod with a new Label and updated data 12/14/22 16:38:11.506 -STEP: getting the Pod and ensuring that it's patched 12/14/22 16:38:11.518 -STEP: replacing the Pod's status Ready condition to False 12/14/22 16:38:11.523 -STEP: check the Pod again to ensure its Ready conditions are False 12/14/22 16:38:11.542 -STEP: deleting the Pod via a Collection with a LabelSelector 12/14/22 16:38:11.543 -STEP: watching for the Pod to be deleted 12/14/22 16:38:11.562 -Dec 14 16:38:11.564: INFO: observed event type MODIFIED -Dec 14 16:38:13.505: INFO: observed event type MODIFIED -Dec 14 16:38:14.521: INFO: observed event type MODIFIED -Dec 14 16:38:14.857: INFO: observed event type MODIFIED -[AfterEach] [sig-node] Pods +STEP: Creating a kubernetes client 02/25/23 10:49:12.216 +Feb 25 10:49:12.216: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename var-expansion 02/25/23 10:49:12.219 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:12.251 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:12.26 +[It] should allow substituting values in a container's command [NodeConformance] [Conformance] + test/e2e/common/node/expansion.go:72 +STEP: Creating a pod to test substitution in container's command 02/25/23 10:49:12.266 +Feb 25 10:49:12.286: INFO: Waiting up to 5m0s for pod "var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971" in namespace "var-expansion-4688" to be "Succeeded or Failed" +Feb 25 10:49:12.303: INFO: Pod "var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971": Phase="Pending", Reason="", readiness=false. Elapsed: 16.943466ms +Feb 25 10:49:14.317: INFO: Pod "var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030688678s +Feb 25 10:49:16.318: INFO: Pod "var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971": Phase="Pending", Reason="", readiness=false. Elapsed: 4.031712769s +Feb 25 10:49:18.311: INFO: Pod "var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971": Phase="Pending", Reason="", readiness=false. Elapsed: 6.025022418s +Feb 25 10:49:20.327: INFO: Pod "var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.040451726s +STEP: Saw pod success 02/25/23 10:49:20.327 +Feb 25 10:49:20.328: INFO: Pod "var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971" satisfied condition "Succeeded or Failed" +Feb 25 10:49:20.344: INFO: Trying to get logs from node diek3dua9gay-3 pod var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971 container dapi-container: +STEP: delete the pod 02/25/23 10:49:20.363 +Feb 25 10:49:20.395: INFO: Waiting for pod var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971 to disappear +Feb 25 10:49:20.419: INFO: Pod var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971 no longer exists +[AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 -Dec 14 16:38:14.866: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-2359" for this suite. 12/14/22 16:38:14.872 -{"msg":"PASSED [sig-node] Pods should run through the lifecycle of Pods and PodStatus [Conformance]","completed":204,"skipped":3806,"failed":0} +Feb 25 10:49:20.419: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "var-expansion-4688" for this suite. 02/25/23 10:49:20.429 +{"msg":"PASSED [sig-node] Variable Expansion should allow substituting values in a container's command [NodeConformance] [Conformance]","completed":202,"skipped":3715,"failed":0} ------------------------------ -• [SLOW TEST] [5.157 seconds] -[sig-node] Pods +• [SLOW TEST] [8.227 seconds] +[sig-node] Variable Expansion test/e2e/common/node/framework.go:23 - should run through the lifecycle of Pods and PodStatus [Conformance] - test/e2e/common/node/pods.go:895 + should allow substituting values in a container's command [NodeConformance] [Conformance] + test/e2e/common/node/expansion.go:72 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Pods + [BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:38:09.725 - Dec 14 16:38:09.725: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename pods 12/14/22 16:38:09.73 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:38:09.755 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:38:09.76 - [BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 - [It] should run through the lifecycle of Pods and PodStatus [Conformance] - test/e2e/common/node/pods.go:895 - STEP: creating a Pod with a static label 12/14/22 16:38:09.772 - STEP: watching for Pod to be ready 12/14/22 16:38:09.787 - Dec 14 16:38:09.789: INFO: observed Pod pod-test in namespace pods-2359 in phase Pending with labels: map[test-pod-static:true] & conditions [] - Dec 14 16:38:09.800: INFO: observed Pod pod-test in namespace pods-2359 in phase Pending with labels: map[test-pod-static:true] & conditions [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:09 +0000 UTC }] - Dec 14 16:38:10.224: INFO: observed Pod pod-test in namespace pods-2359 in phase Pending with labels: map[test-pod-static:true] & conditions [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:09 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:09 +0000 UTC ContainersNotReady containers with unready status: [pod-test]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:09 +0000 UTC ContainersNotReady containers with unready status: [pod-test]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:09 +0000 UTC }] - Dec 14 16:38:11.500: INFO: Found Pod pod-test in namespace pods-2359 in phase Running with labels: map[test-pod-static:true] & conditions [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:09 +0000 UTC } {Ready True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:11 +0000 UTC } {ContainersReady True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:11 +0000 UTC } {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2022-12-14 16:38:09 +0000 UTC }] - STEP: patching the Pod with a new Label and updated data 12/14/22 16:38:11.506 - STEP: getting the Pod and ensuring that it's patched 12/14/22 16:38:11.518 - STEP: replacing the Pod's status Ready condition to False 12/14/22 16:38:11.523 - STEP: check the Pod again to ensure its Ready conditions are False 12/14/22 16:38:11.542 - STEP: deleting the Pod via a Collection with a LabelSelector 12/14/22 16:38:11.543 - STEP: watching for the Pod to be deleted 12/14/22 16:38:11.562 - Dec 14 16:38:11.564: INFO: observed event type MODIFIED - Dec 14 16:38:13.505: INFO: observed event type MODIFIED - Dec 14 16:38:14.521: INFO: observed event type MODIFIED - Dec 14 16:38:14.857: INFO: observed event type MODIFIED - [AfterEach] [sig-node] Pods + STEP: Creating a kubernetes client 02/25/23 10:49:12.216 + Feb 25 10:49:12.216: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename var-expansion 02/25/23 10:49:12.219 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:12.251 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:12.26 + [It] should allow substituting values in a container's command [NodeConformance] [Conformance] + test/e2e/common/node/expansion.go:72 + STEP: Creating a pod to test substitution in container's command 02/25/23 10:49:12.266 + Feb 25 10:49:12.286: INFO: Waiting up to 5m0s for pod "var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971" in namespace "var-expansion-4688" to be "Succeeded or Failed" + Feb 25 10:49:12.303: INFO: Pod "var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971": Phase="Pending", Reason="", readiness=false. Elapsed: 16.943466ms + Feb 25 10:49:14.317: INFO: Pod "var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030688678s + Feb 25 10:49:16.318: INFO: Pod "var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971": Phase="Pending", Reason="", readiness=false. Elapsed: 4.031712769s + Feb 25 10:49:18.311: INFO: Pod "var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971": Phase="Pending", Reason="", readiness=false. Elapsed: 6.025022418s + Feb 25 10:49:20.327: INFO: Pod "var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.040451726s + STEP: Saw pod success 02/25/23 10:49:20.327 + Feb 25 10:49:20.328: INFO: Pod "var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971" satisfied condition "Succeeded or Failed" + Feb 25 10:49:20.344: INFO: Trying to get logs from node diek3dua9gay-3 pod var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971 container dapi-container: + STEP: delete the pod 02/25/23 10:49:20.363 + Feb 25 10:49:20.395: INFO: Waiting for pod var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971 to disappear + Feb 25 10:49:20.419: INFO: Pod var-expansion-683643fc-6809-4aa9-a83c-32d7b8589971 no longer exists + [AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 - Dec 14 16:38:14.866: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "pods-2359" for this suite. 12/14/22 16:38:14.872 + Feb 25 10:49:20.419: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "var-expansion-4688" for this suite. 02/25/23 10:49:20.429 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - should honor timeout [Conformance] - test/e2e/apimachinery/webhook.go:380 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +[sig-auth] ServiceAccounts + should mount an API token into pods [Conformance] + test/e2e/auth/service_accounts.go:75 +[BeforeEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:38:14.898 -Dec 14 16:38:14.898: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 16:38:14.901 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:38:14.934 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:38:14.939 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 16:38:14.958 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:38:16.297 -STEP: Deploying the webhook pod 12/14/22 16:38:16.315 -STEP: Wait for the deployment to be ready 12/14/22 16:38:16.328 -Dec 14 16:38:16.335: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created -STEP: Deploying the webhook service 12/14/22 16:38:18.353 -STEP: Verifying the service has paired with the endpoint 12/14/22 16:38:18.371 -Dec 14 16:38:19.371: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] should honor timeout [Conformance] - test/e2e/apimachinery/webhook.go:380 -STEP: Setting timeout (1s) shorter than webhook latency (5s) 12/14/22 16:38:19.375 -STEP: Registering slow webhook via the AdmissionRegistration API 12/14/22 16:38:19.375 -STEP: Request fails when timeout (1s) is shorter than slow webhook latency (5s) 12/14/22 16:38:19.396 -STEP: Having no error when timeout is shorter than webhook latency and failure policy is ignore 12/14/22 16:38:20.412 -STEP: Registering slow webhook via the AdmissionRegistration API 12/14/22 16:38:20.412 -Dec 14 16:38:20.432: INFO: Waiting for webhook configuration to be ready... -STEP: Having no error when timeout is longer than webhook latency 12/14/22 16:38:21.572 -STEP: Registering slow webhook via the AdmissionRegistration API 12/14/22 16:38:21.572 -STEP: Having no error when timeout is empty (defaulted to 10s in v1) 12/14/22 16:38:26.618 -STEP: Registering slow webhook via the AdmissionRegistration API 12/14/22 16:38:26.618 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 10:49:20.447 +Feb 25 10:49:20.447: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename svcaccounts 02/25/23 10:49:20.451 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:20.498 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:20.515 +[It] should mount an API token into pods [Conformance] + test/e2e/auth/service_accounts.go:75 +Feb 25 10:49:20.552: INFO: Waiting up to 5m0s for pod "pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686" in namespace "svcaccounts-2214" to be "running" +Feb 25 10:49:20.563: INFO: Pod "pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686": Phase="Pending", Reason="", readiness=false. Elapsed: 10.756307ms +Feb 25 10:49:22.577: INFO: Pod "pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025446956s +Feb 25 10:49:24.571: INFO: Pod "pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686": Phase="Running", Reason="", readiness=true. Elapsed: 4.019434601s +Feb 25 10:49:24.571: INFO: Pod "pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686" satisfied condition "running" +STEP: reading a file in the container 02/25/23 10:49:24.571 +Feb 25 10:49:24.583: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-2214 pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token' +STEP: reading a file in the container 02/25/23 10:49:25.132 +Feb 25 10:49:25.134: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-2214 pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt' +STEP: reading a file in the container 02/25/23 10:49:25.763 +Feb 25 10:49:25.764: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-2214 pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace' +Feb 25 10:49:26.543: INFO: Got root ca configmap in namespace "svcaccounts-2214" +[AfterEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:187 -Dec 14 16:38:31.664: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-1741" for this suite. 12/14/22 16:38:31.671 -STEP: Destroying namespace "webhook-1741-markers" for this suite. 12/14/22 16:38:31.682 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should honor timeout [Conformance]","completed":205,"skipped":3836,"failed":0} +Feb 25 10:49:26.548: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "svcaccounts-2214" for this suite. 02/25/23 10:49:26.566 +{"msg":"PASSED [sig-auth] ServiceAccounts should mount an API token into pods [Conformance]","completed":203,"skipped":3766,"failed":0} ------------------------------ -• [SLOW TEST] [16.854 seconds] -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - should honor timeout [Conformance] - test/e2e/apimachinery/webhook.go:380 +• [SLOW TEST] [6.143 seconds] +[sig-auth] ServiceAccounts +test/e2e/auth/framework.go:23 + should mount an API token into pods [Conformance] + test/e2e/auth/service_accounts.go:75 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + [BeforeEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:38:14.898 - Dec 14 16:38:14.898: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 16:38:14.901 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:38:14.934 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:38:14.939 - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 16:38:14.958 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:38:16.297 - STEP: Deploying the webhook pod 12/14/22 16:38:16.315 - STEP: Wait for the deployment to be ready 12/14/22 16:38:16.328 - Dec 14 16:38:16.335: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created - STEP: Deploying the webhook service 12/14/22 16:38:18.353 - STEP: Verifying the service has paired with the endpoint 12/14/22 16:38:18.371 - Dec 14 16:38:19.371: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] should honor timeout [Conformance] - test/e2e/apimachinery/webhook.go:380 - STEP: Setting timeout (1s) shorter than webhook latency (5s) 12/14/22 16:38:19.375 - STEP: Registering slow webhook via the AdmissionRegistration API 12/14/22 16:38:19.375 - STEP: Request fails when timeout (1s) is shorter than slow webhook latency (5s) 12/14/22 16:38:19.396 - STEP: Having no error when timeout is shorter than webhook latency and failure policy is ignore 12/14/22 16:38:20.412 - STEP: Registering slow webhook via the AdmissionRegistration API 12/14/22 16:38:20.412 - Dec 14 16:38:20.432: INFO: Waiting for webhook configuration to be ready... - STEP: Having no error when timeout is longer than webhook latency 12/14/22 16:38:21.572 - STEP: Registering slow webhook via the AdmissionRegistration API 12/14/22 16:38:21.572 - STEP: Having no error when timeout is empty (defaulted to 10s in v1) 12/14/22 16:38:26.618 - STEP: Registering slow webhook via the AdmissionRegistration API 12/14/22 16:38:26.618 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 10:49:20.447 + Feb 25 10:49:20.447: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename svcaccounts 02/25/23 10:49:20.451 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:20.498 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:20.515 + [It] should mount an API token into pods [Conformance] + test/e2e/auth/service_accounts.go:75 + Feb 25 10:49:20.552: INFO: Waiting up to 5m0s for pod "pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686" in namespace "svcaccounts-2214" to be "running" + Feb 25 10:49:20.563: INFO: Pod "pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686": Phase="Pending", Reason="", readiness=false. Elapsed: 10.756307ms + Feb 25 10:49:22.577: INFO: Pod "pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025446956s + Feb 25 10:49:24.571: INFO: Pod "pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686": Phase="Running", Reason="", readiness=true. Elapsed: 4.019434601s + Feb 25 10:49:24.571: INFO: Pod "pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686" satisfied condition "running" + STEP: reading a file in the container 02/25/23 10:49:24.571 + Feb 25 10:49:24.583: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-2214 pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/token' + STEP: reading a file in the container 02/25/23 10:49:25.132 + Feb 25 10:49:25.134: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-2214 pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/ca.crt' + STEP: reading a file in the container 02/25/23 10:49:25.763 + Feb 25 10:49:25.764: INFO: Running '/usr/local/bin/kubectl exec --namespace=svcaccounts-2214 pod-service-account-5e67d3a2-0f6a-4285-ac9b-47e2a6298686 -c=test -- cat /var/run/secrets/kubernetes.io/serviceaccount/namespace' + Feb 25 10:49:26.543: INFO: Got root ca configmap in namespace "svcaccounts-2214" + [AfterEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:187 - Dec 14 16:38:31.664: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-1741" for this suite. 12/14/22 16:38:31.671 - STEP: Destroying namespace "webhook-1741-markers" for this suite. 12/14/22 16:38:31.682 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 + Feb 25 10:49:26.548: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "svcaccounts-2214" for this suite. 02/25/23 10:49:26.566 << End Captured GinkgoWriter Output ------------------------------ -SSSSS +SSSSSSSSSSSSSSSSSS ------------------------------ -[sig-network] Services - should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance] - test/e2e/network/service.go:2157 -[BeforeEach] [sig-network] Services +[sig-apps] ReplicationController + should release no longer matching pods [Conformance] + test/e2e/apps/rc.go:100 +[BeforeEach] [sig-apps] ReplicationController test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:38:31.756 -Dec 14 16:38:31.756: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 16:38:31.759 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:38:31.83 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:38:31.836 -[BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 -[It] should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance] - test/e2e/network/service.go:2157 -STEP: creating service in namespace services-1688 12/14/22 16:38:31.84 -STEP: creating service affinity-clusterip in namespace services-1688 12/14/22 16:38:31.841 -STEP: creating replication controller affinity-clusterip in namespace services-1688 12/14/22 16:38:31.863 -I1214 16:38:31.878376 14 runners.go:193] Created replication controller with name: affinity-clusterip, namespace: services-1688, replica count: 3 -I1214 16:38:34.929604 14 runners.go:193] affinity-clusterip Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -Dec 14 16:38:34.943: INFO: Creating new exec pod -Dec 14 16:38:34.955: INFO: Waiting up to 5m0s for pod "execpod-affinity7xmtg" in namespace "services-1688" to be "running" -Dec 14 16:38:34.959: INFO: Pod "execpod-affinity7xmtg": Phase="Pending", Reason="", readiness=false. Elapsed: 4.106063ms -Dec 14 16:38:36.964: INFO: Pod "execpod-affinity7xmtg": Phase="Running", Reason="", readiness=true. Elapsed: 2.009421062s -Dec 14 16:38:36.964: INFO: Pod "execpod-affinity7xmtg" satisfied condition "running" -Dec 14 16:38:37.966: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1688 exec execpod-affinity7xmtg -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-clusterip 80' -Dec 14 16:38:38.300: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-clusterip 80\nConnection to affinity-clusterip 80 port [tcp/http] succeeded!\n" -Dec 14 16:38:38.300: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:38:38.301: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1688 exec execpod-affinity7xmtg -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.49.96 80' -Dec 14 16:38:38.564: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.49.96 80\nConnection to 10.233.49.96 80 port [tcp/http] succeeded!\n" -Dec 14 16:38:38.564: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:38:38.565: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1688 exec execpod-affinity7xmtg -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.49.96:80/ ; done' -Dec 14 16:38:39.040: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n" -Dec 14 16:38:39.040: INFO: stdout: "\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc" -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc -Dec 14 16:38:39.040: INFO: Cleaning up the exec pod -STEP: deleting ReplicationController affinity-clusterip in namespace services-1688, will wait for the garbage collector to delete the pods 12/14/22 16:38:39.061 -Dec 14 16:38:39.138: INFO: Deleting ReplicationController affinity-clusterip took: 10.669618ms -Dec 14 16:38:39.239: INFO: Terminating ReplicationController affinity-clusterip pods took: 101.377378ms -[AfterEach] [sig-network] Services +STEP: Creating a kubernetes client 02/25/23 10:49:26.593 +Feb 25 10:49:26.593: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename replication-controller 02/25/23 10:49:26.6 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:26.637 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:26.647 +[BeforeEach] [sig-apps] ReplicationController + test/e2e/apps/rc.go:56 +[It] should release no longer matching pods [Conformance] + test/e2e/apps/rc.go:100 +STEP: Given a ReplicationController is created 02/25/23 10:49:26.654 +STEP: When the matched label of one of its pods change 02/25/23 10:49:26.668 +Feb 25 10:49:26.678: INFO: Pod name pod-release: Found 0 pods out of 1 +Feb 25 10:49:31.689: INFO: Pod name pod-release: Found 1 pods out of 1 +STEP: Then the pod is released 02/25/23 10:49:31.715 +[AfterEach] [sig-apps] ReplicationController test/e2e/framework/framework.go:187 -Dec 14 16:38:41.090: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-1688" for this suite. 12/14/22 16:38:41.104 -[AfterEach] [sig-network] Services - test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance]","completed":206,"skipped":3841,"failed":0} +Feb 25 10:49:32.754: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "replication-controller-5232" for this suite. 02/25/23 10:49:32.785 +{"msg":"PASSED [sig-apps] ReplicationController should release no longer matching pods [Conformance]","completed":204,"skipped":3784,"failed":0} ------------------------------ -• [SLOW TEST] [9.357 seconds] -[sig-network] Services -test/e2e/network/common/framework.go:23 - should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance] - test/e2e/network/service.go:2157 +• [SLOW TEST] [6.241 seconds] +[sig-apps] ReplicationController +test/e2e/apps/framework.go:23 + should release no longer matching pods [Conformance] + test/e2e/apps/rc.go:100 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Services + [BeforeEach] [sig-apps] ReplicationController test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:38:31.756 - Dec 14 16:38:31.756: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 16:38:31.759 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:38:31.83 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:38:31.836 - [BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 - [It] should have session affinity work for service with type clusterIP [LinuxOnly] [Conformance] - test/e2e/network/service.go:2157 - STEP: creating service in namespace services-1688 12/14/22 16:38:31.84 - STEP: creating service affinity-clusterip in namespace services-1688 12/14/22 16:38:31.841 - STEP: creating replication controller affinity-clusterip in namespace services-1688 12/14/22 16:38:31.863 - I1214 16:38:31.878376 14 runners.go:193] Created replication controller with name: affinity-clusterip, namespace: services-1688, replica count: 3 - I1214 16:38:34.929604 14 runners.go:193] affinity-clusterip Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - Dec 14 16:38:34.943: INFO: Creating new exec pod - Dec 14 16:38:34.955: INFO: Waiting up to 5m0s for pod "execpod-affinity7xmtg" in namespace "services-1688" to be "running" - Dec 14 16:38:34.959: INFO: Pod "execpod-affinity7xmtg": Phase="Pending", Reason="", readiness=false. Elapsed: 4.106063ms - Dec 14 16:38:36.964: INFO: Pod "execpod-affinity7xmtg": Phase="Running", Reason="", readiness=true. Elapsed: 2.009421062s - Dec 14 16:38:36.964: INFO: Pod "execpod-affinity7xmtg" satisfied condition "running" - Dec 14 16:38:37.966: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1688 exec execpod-affinity7xmtg -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-clusterip 80' - Dec 14 16:38:38.300: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-clusterip 80\nConnection to affinity-clusterip 80 port [tcp/http] succeeded!\n" - Dec 14 16:38:38.300: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:38:38.301: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1688 exec execpod-affinity7xmtg -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.49.96 80' - Dec 14 16:38:38.564: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.49.96 80\nConnection to 10.233.49.96 80 port [tcp/http] succeeded!\n" - Dec 14 16:38:38.564: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:38:38.565: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1688 exec execpod-affinity7xmtg -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://10.233.49.96:80/ ; done' - Dec 14 16:38:39.040: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n+ echo\n+ curl -q -s --connect-timeout 2 http://10.233.49.96:80/\n" - Dec 14 16:38:39.040: INFO: stdout: "\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc\naffinity-clusterip-8qrsc" - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Received response from host: affinity-clusterip-8qrsc - Dec 14 16:38:39.040: INFO: Cleaning up the exec pod - STEP: deleting ReplicationController affinity-clusterip in namespace services-1688, will wait for the garbage collector to delete the pods 12/14/22 16:38:39.061 - Dec 14 16:38:39.138: INFO: Deleting ReplicationController affinity-clusterip took: 10.669618ms - Dec 14 16:38:39.239: INFO: Terminating ReplicationController affinity-clusterip pods took: 101.377378ms - [AfterEach] [sig-network] Services + STEP: Creating a kubernetes client 02/25/23 10:49:26.593 + Feb 25 10:49:26.593: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename replication-controller 02/25/23 10:49:26.6 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:26.637 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:26.647 + [BeforeEach] [sig-apps] ReplicationController + test/e2e/apps/rc.go:56 + [It] should release no longer matching pods [Conformance] + test/e2e/apps/rc.go:100 + STEP: Given a ReplicationController is created 02/25/23 10:49:26.654 + STEP: When the matched label of one of its pods change 02/25/23 10:49:26.668 + Feb 25 10:49:26.678: INFO: Pod name pod-release: Found 0 pods out of 1 + Feb 25 10:49:31.689: INFO: Pod name pod-release: Found 1 pods out of 1 + STEP: Then the pod is released 02/25/23 10:49:31.715 + [AfterEach] [sig-apps] ReplicationController test/e2e/framework/framework.go:187 - Dec 14 16:38:41.090: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-1688" for this suite. 12/14/22 16:38:41.104 - [AfterEach] [sig-network] Services - test/e2e/network/service.go:762 + Feb 25 10:49:32.754: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "replication-controller-5232" for this suite. 02/25/23 10:49:32.785 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook - should execute prestop exec hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:114 -[BeforeEach] [sig-node] Container Lifecycle Hook +[sig-storage] Projected downwardAPI + should provide container's memory limit [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:206 +[BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:38:41.12 -Dec 14 16:38:41.121: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-lifecycle-hook 12/14/22 16:38:41.122 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:38:41.145 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:38:41.147 -[BeforeEach] when create a pod with lifecycle hook - test/e2e/common/node/lifecycle_hook.go:55 -STEP: create the container to handle the HTTPGet hook request. 12/14/22 16:38:41.155 -Dec 14 16:38:41.164: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-6112" to be "running and ready" -Dec 14 16:38:41.173: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 8.95871ms -Dec 14 16:38:41.173: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:38:43.183: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 2.019265887s -Dec 14 16:38:43.183: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) -Dec 14 16:38:43.183: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" -[It] should execute prestop exec hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:114 -STEP: create the pod with lifecycle hook 12/14/22 16:38:43.188 -Dec 14 16:38:43.209: INFO: Waiting up to 5m0s for pod "pod-with-prestop-exec-hook" in namespace "container-lifecycle-hook-6112" to be "running and ready" -Dec 14 16:38:43.215: INFO: Pod "pod-with-prestop-exec-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 5.730467ms -Dec 14 16:38:43.215: INFO: The phase of Pod pod-with-prestop-exec-hook is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:38:45.226: INFO: Pod "pod-with-prestop-exec-hook": Phase="Running", Reason="", readiness=true. Elapsed: 2.017435905s -Dec 14 16:38:45.226: INFO: The phase of Pod pod-with-prestop-exec-hook is Running (Ready = true) -Dec 14 16:38:45.226: INFO: Pod "pod-with-prestop-exec-hook" satisfied condition "running and ready" -STEP: delete the pod with lifecycle hook 12/14/22 16:38:45.231 -Dec 14 16:38:45.246: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -Dec 14 16:38:45.253: INFO: Pod pod-with-prestop-exec-hook still exists -Dec 14 16:38:47.254: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -Dec 14 16:38:47.263: INFO: Pod pod-with-prestop-exec-hook still exists -Dec 14 16:38:49.254: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear -Dec 14 16:38:49.262: INFO: Pod pod-with-prestop-exec-hook no longer exists -STEP: check prestop hook 12/14/22 16:38:49.262 -[AfterEach] [sig-node] Container Lifecycle Hook +STEP: Creating a kubernetes client 02/25/23 10:49:32.861 +Feb 25 10:49:32.862: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 10:49:32.868 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:32.999 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:33.013 +[BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 +[It] should provide container's memory limit [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:206 +STEP: Creating a pod to test downward API volume plugin 02/25/23 10:49:33.024 +Feb 25 10:49:33.062: INFO: Waiting up to 5m0s for pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1" in namespace "projected-8012" to be "Succeeded or Failed" +Feb 25 10:49:33.084: INFO: Pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1": Phase="Pending", Reason="", readiness=false. Elapsed: 22.503305ms +Feb 25 10:49:35.094: INFO: Pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031979219s +Feb 25 10:49:37.160: INFO: Pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1": Phase="Pending", Reason="", readiness=false. Elapsed: 4.098574768s +Feb 25 10:49:39.114: INFO: Pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1": Phase="Pending", Reason="", readiness=false. Elapsed: 6.052441058s +Feb 25 10:49:41.093: INFO: Pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1": Phase="Pending", Reason="", readiness=false. Elapsed: 8.031387714s +Feb 25 10:49:43.096: INFO: Pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.034185675s +STEP: Saw pod success 02/25/23 10:49:43.096 +Feb 25 10:49:43.097: INFO: Pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1" satisfied condition "Succeeded or Failed" +Feb 25 10:49:43.104: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1 container client-container: +STEP: delete the pod 02/25/23 10:49:43.136 +Feb 25 10:49:43.174: INFO: Waiting for pod downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1 to disappear +Feb 25 10:49:43.189: INFO: Pod downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1 no longer exists +[AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 -Dec 14 16:38:49.293: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-lifecycle-hook-6112" for this suite. 12/14/22 16:38:49.301 -{"msg":"PASSED [sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop exec hook properly [NodeConformance] [Conformance]","completed":207,"skipped":3875,"failed":0} +Feb 25 10:49:43.189: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-8012" for this suite. 02/25/23 10:49:43.209 +{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's memory limit [NodeConformance] [Conformance]","completed":205,"skipped":3823,"failed":0} ------------------------------ -• [SLOW TEST] [8.193 seconds] -[sig-node] Container Lifecycle Hook -test/e2e/common/node/framework.go:23 - when create a pod with lifecycle hook - test/e2e/common/node/lifecycle_hook.go:46 - should execute prestop exec hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:114 +• [SLOW TEST] [10.375 seconds] +[sig-storage] Projected downwardAPI +test/e2e/common/storage/framework.go:23 + should provide container's memory limit [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:206 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Container Lifecycle Hook + [BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:38:41.12 - Dec 14 16:38:41.121: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-lifecycle-hook 12/14/22 16:38:41.122 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:38:41.145 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:38:41.147 - [BeforeEach] when create a pod with lifecycle hook - test/e2e/common/node/lifecycle_hook.go:55 - STEP: create the container to handle the HTTPGet hook request. 12/14/22 16:38:41.155 - Dec 14 16:38:41.164: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-6112" to be "running and ready" - Dec 14 16:38:41.173: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 8.95871ms - Dec 14 16:38:41.173: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:38:43.183: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 2.019265887s - Dec 14 16:38:43.183: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) - Dec 14 16:38:43.183: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" - [It] should execute prestop exec hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:114 - STEP: create the pod with lifecycle hook 12/14/22 16:38:43.188 - Dec 14 16:38:43.209: INFO: Waiting up to 5m0s for pod "pod-with-prestop-exec-hook" in namespace "container-lifecycle-hook-6112" to be "running and ready" - Dec 14 16:38:43.215: INFO: Pod "pod-with-prestop-exec-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 5.730467ms - Dec 14 16:38:43.215: INFO: The phase of Pod pod-with-prestop-exec-hook is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:38:45.226: INFO: Pod "pod-with-prestop-exec-hook": Phase="Running", Reason="", readiness=true. Elapsed: 2.017435905s - Dec 14 16:38:45.226: INFO: The phase of Pod pod-with-prestop-exec-hook is Running (Ready = true) - Dec 14 16:38:45.226: INFO: Pod "pod-with-prestop-exec-hook" satisfied condition "running and ready" - STEP: delete the pod with lifecycle hook 12/14/22 16:38:45.231 - Dec 14 16:38:45.246: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear - Dec 14 16:38:45.253: INFO: Pod pod-with-prestop-exec-hook still exists - Dec 14 16:38:47.254: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear - Dec 14 16:38:47.263: INFO: Pod pod-with-prestop-exec-hook still exists - Dec 14 16:38:49.254: INFO: Waiting for pod pod-with-prestop-exec-hook to disappear - Dec 14 16:38:49.262: INFO: Pod pod-with-prestop-exec-hook no longer exists - STEP: check prestop hook 12/14/22 16:38:49.262 - [AfterEach] [sig-node] Container Lifecycle Hook + STEP: Creating a kubernetes client 02/25/23 10:49:32.861 + Feb 25 10:49:32.862: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 10:49:32.868 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:32.999 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:33.013 + [BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 + [It] should provide container's memory limit [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:206 + STEP: Creating a pod to test downward API volume plugin 02/25/23 10:49:33.024 + Feb 25 10:49:33.062: INFO: Waiting up to 5m0s for pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1" in namespace "projected-8012" to be "Succeeded or Failed" + Feb 25 10:49:33.084: INFO: Pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1": Phase="Pending", Reason="", readiness=false. Elapsed: 22.503305ms + Feb 25 10:49:35.094: INFO: Pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031979219s + Feb 25 10:49:37.160: INFO: Pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1": Phase="Pending", Reason="", readiness=false. Elapsed: 4.098574768s + Feb 25 10:49:39.114: INFO: Pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1": Phase="Pending", Reason="", readiness=false. Elapsed: 6.052441058s + Feb 25 10:49:41.093: INFO: Pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1": Phase="Pending", Reason="", readiness=false. Elapsed: 8.031387714s + Feb 25 10:49:43.096: INFO: Pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 10.034185675s + STEP: Saw pod success 02/25/23 10:49:43.096 + Feb 25 10:49:43.097: INFO: Pod "downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1" satisfied condition "Succeeded or Failed" + Feb 25 10:49:43.104: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1 container client-container: + STEP: delete the pod 02/25/23 10:49:43.136 + Feb 25 10:49:43.174: INFO: Waiting for pod downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1 to disappear + Feb 25 10:49:43.189: INFO: Pod downwardapi-volume-39ef339c-a660-44de-a4a5-eefd656cddd1 no longer exists + [AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 - Dec 14 16:38:49.293: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-lifecycle-hook-6112" for this suite. 12/14/22 16:38:49.301 + Feb 25 10:49:43.189: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-8012" for this suite. 02/25/23 10:49:43.209 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-apps] Daemon set [Serial] - should rollback without unnecessary restarts [Conformance] - test/e2e/apps/daemon_set.go:431 -[BeforeEach] [sig-apps] Daemon set [Serial] +[sig-instrumentation] Events API + should ensure that an event can be fetched, patched, deleted, and listed [Conformance] + test/e2e/instrumentation/events.go:98 +[BeforeEach] [sig-instrumentation] Events API test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:38:49.321 -Dec 14 16:38:49.321: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename daemonsets 12/14/22 16:38:49.323 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:38:49.347 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:38:49.35 -[BeforeEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:145 -[It] should rollback without unnecessary restarts [Conformance] - test/e2e/apps/daemon_set.go:431 -Dec 14 16:38:49.399: INFO: Create a RollingUpdate DaemonSet -Dec 14 16:38:49.406: INFO: Check that daemon pods launch on every node of the cluster -Dec 14 16:38:49.418: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:38:49.418: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 16:38:50.438: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:38:50.438: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 16:38:51.437: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 -Dec 14 16:38:51.437: INFO: Node kiedai4iakae-2 is running 0 daemon pod, expected 1 -Dec 14 16:38:52.436: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 -Dec 14 16:38:52.436: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set -Dec 14 16:38:52.436: INFO: Update the DaemonSet to trigger a rollout -Dec 14 16:38:52.457: INFO: Updating DaemonSet daemon-set -Dec 14 16:38:55.485: INFO: Roll back the DaemonSet before rollout is complete -Dec 14 16:38:55.514: INFO: Updating DaemonSet daemon-set -Dec 14 16:38:55.514: INFO: Make sure DaemonSet rollback is complete -Dec 14 16:38:55.522: INFO: Wrong image for pod: daemon-set-vvz87. Expected: registry.k8s.io/e2e-test-images/httpd:2.4.38-2, got: foo:non-existent. -Dec 14 16:38:55.522: INFO: Pod daemon-set-vvz87 is not available -Dec 14 16:39:00.535: INFO: Pod daemon-set-r6fv5 is not available -[AfterEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:110 -STEP: Deleting DaemonSet "daemon-set" 12/14/22 16:39:00.55 -STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-8809, will wait for the garbage collector to delete the pods 12/14/22 16:39:00.55 -Dec 14 16:39:00.617: INFO: Deleting DaemonSet.extensions daemon-set took: 9.876809ms -Dec 14 16:39:00.717: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.678867ms -Dec 14 16:39:02.024: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 16:39:02.024: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set -Dec 14 16:39:02.029: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"22049"},"items":null} - -Dec 14 16:39:02.034: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"22049"},"items":null} - -[AfterEach] [sig-apps] Daemon set [Serial] +STEP: Creating a kubernetes client 02/25/23 10:49:43.248 +Feb 25 10:49:43.248: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename events 02/25/23 10:49:43.252 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:43.291 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:43.3 +[BeforeEach] [sig-instrumentation] Events API + test/e2e/instrumentation/events.go:84 +[It] should ensure that an event can be fetched, patched, deleted, and listed [Conformance] + test/e2e/instrumentation/events.go:98 +STEP: creating a test event 02/25/23 10:49:43.313 +STEP: listing events in all namespaces 02/25/23 10:49:43.338 +STEP: listing events in test namespace 02/25/23 10:49:43.367 +STEP: listing events with field selection filtering on source 02/25/23 10:49:43.377 +STEP: listing events with field selection filtering on reportingController 02/25/23 10:49:43.389 +STEP: getting the test event 02/25/23 10:49:43.405 +STEP: patching the test event 02/25/23 10:49:43.462 +STEP: getting the test event 02/25/23 10:49:43.521 +STEP: updating the test event 02/25/23 10:49:43.532 +STEP: getting the test event 02/25/23 10:49:43.551 +STEP: deleting the test event 02/25/23 10:49:43.575 +STEP: listing events in all namespaces 02/25/23 10:49:43.612 +STEP: listing events in test namespace 02/25/23 10:49:43.627 +[AfterEach] [sig-instrumentation] Events API test/e2e/framework/framework.go:187 -Dec 14 16:39:02.066: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "daemonsets-8809" for this suite. 12/14/22 16:39:02.071 -{"msg":"PASSED [sig-apps] Daemon set [Serial] should rollback without unnecessary restarts [Conformance]","completed":208,"skipped":3878,"failed":0} +Feb 25 10:49:43.646: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "events-5389" for this suite. 02/25/23 10:49:43.658 +{"msg":"PASSED [sig-instrumentation] Events API should ensure that an event can be fetched, patched, deleted, and listed [Conformance]","completed":206,"skipped":3850,"failed":0} ------------------------------ -• [SLOW TEST] [12.764 seconds] -[sig-apps] Daemon set [Serial] -test/e2e/apps/framework.go:23 - should rollback without unnecessary restarts [Conformance] - test/e2e/apps/daemon_set.go:431 +• [0.436 seconds] +[sig-instrumentation] Events API +test/e2e/instrumentation/common/framework.go:23 + should ensure that an event can be fetched, patched, deleted, and listed [Conformance] + test/e2e/instrumentation/events.go:98 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Daemon set [Serial] + [BeforeEach] [sig-instrumentation] Events API test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:38:49.321 - Dec 14 16:38:49.321: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename daemonsets 12/14/22 16:38:49.323 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:38:49.347 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:38:49.35 - [BeforeEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:145 - [It] should rollback without unnecessary restarts [Conformance] - test/e2e/apps/daemon_set.go:431 - Dec 14 16:38:49.399: INFO: Create a RollingUpdate DaemonSet - Dec 14 16:38:49.406: INFO: Check that daemon pods launch on every node of the cluster - Dec 14 16:38:49.418: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:38:49.418: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 16:38:50.438: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:38:50.438: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 16:38:51.437: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 - Dec 14 16:38:51.437: INFO: Node kiedai4iakae-2 is running 0 daemon pod, expected 1 - Dec 14 16:38:52.436: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 - Dec 14 16:38:52.436: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set - Dec 14 16:38:52.436: INFO: Update the DaemonSet to trigger a rollout - Dec 14 16:38:52.457: INFO: Updating DaemonSet daemon-set - Dec 14 16:38:55.485: INFO: Roll back the DaemonSet before rollout is complete - Dec 14 16:38:55.514: INFO: Updating DaemonSet daemon-set - Dec 14 16:38:55.514: INFO: Make sure DaemonSet rollback is complete - Dec 14 16:38:55.522: INFO: Wrong image for pod: daemon-set-vvz87. Expected: registry.k8s.io/e2e-test-images/httpd:2.4.38-2, got: foo:non-existent. - Dec 14 16:38:55.522: INFO: Pod daemon-set-vvz87 is not available - Dec 14 16:39:00.535: INFO: Pod daemon-set-r6fv5 is not available - [AfterEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:110 - STEP: Deleting DaemonSet "daemon-set" 12/14/22 16:39:00.55 - STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-8809, will wait for the garbage collector to delete the pods 12/14/22 16:39:00.55 - Dec 14 16:39:00.617: INFO: Deleting DaemonSet.extensions daemon-set took: 9.876809ms - Dec 14 16:39:00.717: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.678867ms - Dec 14 16:39:02.024: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 16:39:02.024: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set - Dec 14 16:39:02.029: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"22049"},"items":null} - - Dec 14 16:39:02.034: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"22049"},"items":null} - - [AfterEach] [sig-apps] Daemon set [Serial] + STEP: Creating a kubernetes client 02/25/23 10:49:43.248 + Feb 25 10:49:43.248: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename events 02/25/23 10:49:43.252 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:43.291 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:43.3 + [BeforeEach] [sig-instrumentation] Events API + test/e2e/instrumentation/events.go:84 + [It] should ensure that an event can be fetched, patched, deleted, and listed [Conformance] + test/e2e/instrumentation/events.go:98 + STEP: creating a test event 02/25/23 10:49:43.313 + STEP: listing events in all namespaces 02/25/23 10:49:43.338 + STEP: listing events in test namespace 02/25/23 10:49:43.367 + STEP: listing events with field selection filtering on source 02/25/23 10:49:43.377 + STEP: listing events with field selection filtering on reportingController 02/25/23 10:49:43.389 + STEP: getting the test event 02/25/23 10:49:43.405 + STEP: patching the test event 02/25/23 10:49:43.462 + STEP: getting the test event 02/25/23 10:49:43.521 + STEP: updating the test event 02/25/23 10:49:43.532 + STEP: getting the test event 02/25/23 10:49:43.551 + STEP: deleting the test event 02/25/23 10:49:43.575 + STEP: listing events in all namespaces 02/25/23 10:49:43.612 + STEP: listing events in test namespace 02/25/23 10:49:43.627 + [AfterEach] [sig-instrumentation] Events API test/e2e/framework/framework.go:187 - Dec 14 16:39:02.066: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "daemonsets-8809" for this suite. 12/14/22 16:39:02.071 + Feb 25 10:49:43.646: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "events-5389" for this suite. 02/25/23 10:49:43.658 << End Captured GinkgoWriter Output ------------------------------ -SSS +S ------------------------------ -[sig-apps] DisruptionController - should update/patch PodDisruptionBudget status [Conformance] - test/e2e/apps/disruption.go:163 -[BeforeEach] [sig-apps] DisruptionController +[sig-scheduling] SchedulerPredicates [Serial] + validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] + test/e2e/scheduling/predicates.go:699 +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:39:02.086 -Dec 14 16:39:02.086: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename disruption 12/14/22 16:39:02.089 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:39:02.11 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:39:02.114 -[BeforeEach] [sig-apps] DisruptionController - test/e2e/apps/disruption.go:71 -[It] should update/patch PodDisruptionBudget status [Conformance] - test/e2e/apps/disruption.go:163 -STEP: Waiting for the pdb to be processed 12/14/22 16:39:02.124 -STEP: Updating PodDisruptionBudget status 12/14/22 16:39:04.14 -STEP: Waiting for all pods to be running 12/14/22 16:39:04.182 -Dec 14 16:39:04.195: INFO: running pods: 0 < 1 -STEP: locating a running pod 12/14/22 16:39:06.203 -STEP: Waiting for the pdb to be processed 12/14/22 16:39:06.221 -STEP: Patching PodDisruptionBudget status 12/14/22 16:39:06.231 -STEP: Waiting for the pdb to be processed 12/14/22 16:39:06.248 -[AfterEach] [sig-apps] DisruptionController +STEP: Creating a kubernetes client 02/25/23 10:49:43.684 +Feb 25 10:49:43.684: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename sched-pred 02/25/23 10:49:43.688 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:43.774 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:43.781 +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:92 +Feb 25 10:49:43.786: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready +Feb 25 10:49:43.807: INFO: Waiting for terminating namespaces to be deleted... +Feb 25 10:49:43.816: INFO: +Logging pods the apiserver thinks is on node diek3dua9gay-1 before test +Feb 25 10:49:43.844: INFO: echo-other-node-cd69fcf6b-hsnk9 from cilium-test started at 2023-02-25 09:38:06 +0000 UTC (2 container statuses recorded) +Feb 25 10:49:43.844: INFO: Container dns-test-server ready: true, restart count 0 +Feb 25 10:49:43.844: INFO: Container echo-other-node ready: true, restart count 0 +Feb 25 10:49:43.844: INFO: cilium-node-init-s84z5 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.844: INFO: Container node-init ready: true, restart count 0 +Feb 25 10:49:43.844: INFO: cilium-r4lc4 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.844: INFO: Container cilium-agent ready: true, restart count 0 +Feb 25 10:49:43.844: INFO: coredns-565d847f94-qqvf6 from kube-system started at 2023-02-25 08:49:06 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.844: INFO: Container coredns ready: true, restart count 0 +Feb 25 10:49:43.845: INFO: kube-addon-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.845: INFO: Container kube-addon-manager ready: true, restart count 0 +Feb 25 10:49:43.845: INFO: kube-apiserver-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:09 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.845: INFO: Container kube-apiserver ready: true, restart count 0 +Feb 25 10:49:43.845: INFO: kube-controller-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:49:55 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.845: INFO: Container kube-controller-manager ready: true, restart count 0 +Feb 25 10:49:43.845: INFO: kube-proxy-cm4qg from kube-system started at 2023-02-25 08:46:38 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.845: INFO: Container kube-proxy ready: true, restart count 0 +Feb 25 10:49:43.845: INFO: kube-scheduler-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:27 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.845: INFO: Container kube-scheduler ready: true, restart count 0 +Feb 25 10:49:43.845: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-wznrx from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 10:49:43.846: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 10:49:43.846: INFO: Container systemd-logs ready: true, restart count 0 +Feb 25 10:49:43.846: INFO: +Logging pods the apiserver thinks is on node diek3dua9gay-2 before test +Feb 25 10:49:43.907: INFO: cilium-9tkfg from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.908: INFO: Container cilium-agent ready: true, restart count 0 +Feb 25 10:49:43.908: INFO: cilium-node-init-8b7n2 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.908: INFO: Container node-init ready: true, restart count 0 +Feb 25 10:49:43.908: INFO: cilium-operator-59d7d77cdc-lppbd from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.908: INFO: Container cilium-operator ready: true, restart count 0 +Feb 25 10:49:43.908: INFO: coredns-565d847f94-5nh2n from kube-system started at 2023-02-25 08:48:51 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.908: INFO: Container coredns ready: true, restart count 0 +Feb 25 10:49:43.908: INFO: kube-addon-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.908: INFO: Container kube-addon-manager ready: true, restart count 0 +Feb 25 10:49:43.908: INFO: kube-apiserver-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.908: INFO: Container kube-apiserver ready: true, restart count 0 +Feb 25 10:49:43.908: INFO: kube-controller-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.908: INFO: Container kube-controller-manager ready: true, restart count 0 +Feb 25 10:49:43.908: INFO: kube-proxy-d66m7 from kube-system started at 2023-02-25 08:46:57 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.909: INFO: Container kube-proxy ready: true, restart count 0 +Feb 25 10:49:43.909: INFO: kube-scheduler-diek3dua9gay-2 from kube-system started at 2023-02-25 08:50:15 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.909: INFO: Container kube-scheduler ready: true, restart count 0 +Feb 25 10:49:43.909: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-mml8w from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 10:49:43.909: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 10:49:43.909: INFO: Container systemd-logs ready: true, restart count 0 +Feb 25 10:49:43.909: INFO: +Logging pods the apiserver thinks is on node diek3dua9gay-3 before test +Feb 25 10:49:43.955: INFO: client-7b78db77d5-27xbt from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.955: INFO: Container client ready: true, restart count 0 +Feb 25 10:49:43.955: INFO: client2-78f748dd67-skwz2 from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.955: INFO: Container client2 ready: true, restart count 0 +Feb 25 10:49:43.955: INFO: echo-same-node-85bc9b6b56-wqmfq from cilium-test started at 2023-02-25 09:57:46 +0000 UTC (2 container statuses recorded) +Feb 25 10:49:43.955: INFO: Container dns-test-server ready: true, restart count 0 +Feb 25 10:49:43.955: INFO: Container echo-same-node ready: true, restart count 0 +Feb 25 10:49:43.955: INFO: cilium-2cncz from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.955: INFO: Container cilium-agent ready: true, restart count 0 +Feb 25 10:49:43.955: INFO: cilium-node-init-979mb from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.955: INFO: Container node-init ready: true, restart count 0 +Feb 25 10:49:43.955: INFO: kube-proxy-22s9q from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.955: INFO: Container kube-proxy ready: true, restart count 0 +Feb 25 10:49:43.955: INFO: sonobuoy from sonobuoy started at 2023-02-25 09:45:14 +0000 UTC (1 container statuses recorded) +Feb 25 10:49:43.955: INFO: Container kube-sonobuoy ready: true, restart count 0 +Feb 25 10:49:43.955: INFO: sonobuoy-e2e-job-0b3c30975b5c4880 from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 10:49:43.955: INFO: Container e2e ready: true, restart count 0 +Feb 25 10:49:43.955: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 10:49:43.955: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-9pdjb from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 10:49:43.955: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 10:49:43.955: INFO: Container systemd-logs ready: true, restart count 0 +[It] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] + test/e2e/scheduling/predicates.go:699 +STEP: Trying to launch a pod without a label to get a node which can launch it. 02/25/23 10:49:43.955 +Feb 25 10:49:43.983: INFO: Waiting up to 1m0s for pod "without-label" in namespace "sched-pred-7417" to be "running" +Feb 25 10:49:43.999: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 15.815522ms +Feb 25 10:49:46.015: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031180874s +Feb 25 10:49:48.009: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 4.025307926s +Feb 25 10:49:50.008: INFO: Pod "without-label": Phase="Running", Reason="", readiness=true. Elapsed: 6.024639364s +Feb 25 10:49:50.008: INFO: Pod "without-label" satisfied condition "running" +STEP: Explicitly delete pod here to free the resource it takes. 02/25/23 10:49:50.022 +STEP: Trying to apply a random label on the found node. 02/25/23 10:49:50.069 +STEP: verifying the node has the label kubernetes.io/e2e-cd126814-d2ad-4e65-b03d-da1668bbdba1 95 02/25/23 10:49:50.146 +STEP: Trying to create a pod(pod4) with hostport 54322 and hostIP 0.0.0.0(empty string here) and expect scheduled 02/25/23 10:49:50.194 +Feb 25 10:49:50.219: INFO: Waiting up to 5m0s for pod "pod4" in namespace "sched-pred-7417" to be "not pending" +Feb 25 10:49:50.244: INFO: Pod "pod4": Phase="Pending", Reason="", readiness=false. Elapsed: 24.867341ms +Feb 25 10:49:52.254: INFO: Pod "pod4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.0355768s +Feb 25 10:49:54.255: INFO: Pod "pod4": Phase="Pending", Reason="", readiness=false. Elapsed: 4.035668718s +Feb 25 10:49:56.254: INFO: Pod "pod4": Phase="Running", Reason="", readiness=false. Elapsed: 6.035123734s +Feb 25 10:49:56.254: INFO: Pod "pod4" satisfied condition "not pending" +STEP: Trying to create another pod(pod5) with hostport 54322 but hostIP 192.168.121.44 on the node which pod4 resides and expect not scheduled 02/25/23 10:49:56.254 +Feb 25 10:49:56.267: INFO: Waiting up to 5m0s for pod "pod5" in namespace "sched-pred-7417" to be "not pending" +Feb 25 10:49:56.277: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 10.372375ms +Feb 25 10:49:58.285: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017754162s +Feb 25 10:50:00.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4.022980694s +Feb 25 10:50:02.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 6.018774556s +Feb 25 10:50:04.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 8.019494425s +Feb 25 10:50:06.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 10.023413684s +Feb 25 10:50:08.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 12.021006876s +Feb 25 10:50:10.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 14.021083628s +Feb 25 10:50:12.291: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 16.024000127s +Feb 25 10:50:14.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 18.021954878s +Feb 25 10:50:16.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 20.019869062s +Feb 25 10:50:18.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 22.019178394s +Feb 25 10:50:20.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 24.018806555s +Feb 25 10:50:22.295: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 26.027456039s +Feb 25 10:50:24.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 28.020874362s +Feb 25 10:50:26.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 30.019443817s +Feb 25 10:50:28.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 32.020507907s +Feb 25 10:50:30.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 34.020302952s +Feb 25 10:50:32.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 36.023157663s +Feb 25 10:50:34.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 38.020646692s +Feb 25 10:50:36.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 40.020758481s +Feb 25 10:50:38.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 42.021574844s +Feb 25 10:50:40.296: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 44.028637513s +Feb 25 10:50:42.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 46.021122162s +Feb 25 10:50:44.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 48.01910861s +Feb 25 10:50:46.292: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 50.024741857s +Feb 25 10:50:48.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 52.021488238s +Feb 25 10:50:50.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 54.019052153s +Feb 25 10:50:52.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 56.02222613s +Feb 25 10:50:54.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 58.023262639s +Feb 25 10:50:56.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m0.021449895s +Feb 25 10:50:58.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m2.022250317s +Feb 25 10:51:00.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m4.020838143s +Feb 25 10:51:02.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m6.019857203s +Feb 25 10:51:04.299: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m8.031765945s +Feb 25 10:51:06.285: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m10.01751915s +Feb 25 10:51:08.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m12.021394597s +Feb 25 10:51:10.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m14.01854391s +Feb 25 10:51:12.285: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m16.018260224s +Feb 25 10:51:14.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m18.021428968s +Feb 25 10:51:16.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m20.020308355s +Feb 25 10:51:18.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m22.019923977s +Feb 25 10:51:20.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m24.020542738s +Feb 25 10:51:22.294: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m26.026754742s +Feb 25 10:51:24.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m28.020136091s +Feb 25 10:51:26.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m30.018950333s +Feb 25 10:51:28.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m32.019143054s +Feb 25 10:51:30.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m34.020616866s +Feb 25 10:51:32.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m36.02041168s +Feb 25 10:51:34.300: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m38.032624749s +Feb 25 10:51:36.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m40.020896983s +Feb 25 10:51:38.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m42.0187815s +Feb 25 10:51:40.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m44.019927403s +Feb 25 10:51:42.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m46.021048371s +Feb 25 10:51:44.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m48.02098516s +Feb 25 10:51:46.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m50.021728367s +Feb 25 10:51:48.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m52.01898498s +Feb 25 10:51:50.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m54.019502585s +Feb 25 10:51:52.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m56.018917293s +Feb 25 10:51:54.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m58.019156606s +Feb 25 10:51:56.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m0.019696432s +Feb 25 10:51:58.284: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m2.017294273s +Feb 25 10:52:00.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m4.022107645s +Feb 25 10:52:02.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m6.021368374s +Feb 25 10:52:04.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m8.02255758s +Feb 25 10:52:06.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m10.021543676s +Feb 25 10:52:08.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m12.020435477s +Feb 25 10:52:10.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m14.020006886s +Feb 25 10:52:12.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m16.019260452s +Feb 25 10:52:14.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m18.020325224s +Feb 25 10:52:16.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m20.018903392s +Feb 25 10:52:18.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m22.021393285s +Feb 25 10:52:20.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m24.019728092s +Feb 25 10:52:22.285: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m26.018243995s +Feb 25 10:52:24.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m28.019446298s +Feb 25 10:52:26.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m30.019156597s +Feb 25 10:52:28.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m32.019447136s +Feb 25 10:52:30.291: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m34.023914823s +Feb 25 10:52:32.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m36.018574634s +Feb 25 10:52:34.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m38.021058259s +Feb 25 10:52:36.291: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m40.023578732s +Feb 25 10:52:38.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m42.020325603s +Feb 25 10:52:40.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m44.019960249s +Feb 25 10:52:42.292: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m46.025201232s +Feb 25 10:52:44.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m48.020318496s +Feb 25 10:52:46.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m50.019080334s +Feb 25 10:52:48.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m52.018685964s +Feb 25 10:52:50.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m54.021673044s +Feb 25 10:52:52.293: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m56.026317461s +Feb 25 10:52:54.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m58.022281364s +Feb 25 10:52:56.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m0.019686908s +Feb 25 10:52:58.291: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m2.023920969s +Feb 25 10:53:00.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m4.01973143s +Feb 25 10:53:02.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m6.020784807s +Feb 25 10:53:04.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m8.021606104s +Feb 25 10:53:06.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m10.018510495s +Feb 25 10:53:08.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m12.019752344s +Feb 25 10:53:10.284: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m14.016683863s +Feb 25 10:53:12.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m16.019354017s +Feb 25 10:53:14.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m18.022577365s +Feb 25 10:53:16.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m20.022692292s +Feb 25 10:53:18.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m22.020882238s +Feb 25 10:53:20.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m24.018833112s +Feb 25 10:53:22.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m26.019092518s +Feb 25 10:53:24.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m28.019855317s +Feb 25 10:53:26.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m30.019760913s +Feb 25 10:53:28.285: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m32.017764967s +Feb 25 10:53:30.285: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m34.017666813s +Feb 25 10:53:32.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m36.019768382s +Feb 25 10:53:34.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m38.018762385s +Feb 25 10:53:36.297: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m40.029841512s +Feb 25 10:53:38.285: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m42.018399347s +Feb 25 10:53:40.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m44.020382494s +Feb 25 10:53:42.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m46.018972935s +Feb 25 10:53:44.323: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m48.055950135s +Feb 25 10:53:46.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m50.021043859s +Feb 25 10:53:48.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m52.020673168s +Feb 25 10:53:50.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m54.019314145s +Feb 25 10:53:52.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m56.019043667s +Feb 25 10:53:54.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m58.020658765s +Feb 25 10:53:56.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m0.020677455s +Feb 25 10:53:58.284: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m2.017354685s +Feb 25 10:54:00.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m4.021113242s +Feb 25 10:54:02.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m6.020427842s +Feb 25 10:54:04.292: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m8.024882749s +Feb 25 10:54:06.295: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m10.028206471s +Feb 25 10:54:08.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m12.020606441s +Feb 25 10:54:10.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m14.021653659s +Feb 25 10:54:12.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m16.02263428s +Feb 25 10:54:14.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m18.019703961s +Feb 25 10:54:16.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m20.020184764s +Feb 25 10:54:18.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m22.021723033s +Feb 25 10:54:20.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m24.02017799s +Feb 25 10:54:22.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m26.020964983s +Feb 25 10:54:24.292: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m28.025303638s +Feb 25 10:54:26.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m30.020061323s +Feb 25 10:54:28.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m32.018710104s +Feb 25 10:54:30.291: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m34.023600464s +Feb 25 10:54:32.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m36.019418233s +Feb 25 10:54:34.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m38.019054754s +Feb 25 10:54:36.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m40.018725241s +Feb 25 10:54:38.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m42.020555064s +Feb 25 10:54:40.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m44.019730584s +Feb 25 10:54:42.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m46.020168718s +Feb 25 10:54:44.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m48.020910014s +Feb 25 10:54:46.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m50.018767923s +Feb 25 10:54:48.291: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m52.024337836s +Feb 25 10:54:50.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m54.021607736s +Feb 25 10:54:52.301: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m56.03390871s +Feb 25 10:54:54.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m58.022824183s +Feb 25 10:54:56.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 5m0.019991881s +Feb 25 10:54:56.296: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 5m0.028573083s +STEP: removing the label kubernetes.io/e2e-cd126814-d2ad-4e65-b03d-da1668bbdba1 off the node diek3dua9gay-3 02/25/23 10:54:56.296 +STEP: verifying the node doesn't have the label kubernetes.io/e2e-cd126814-d2ad-4e65-b03d-da1668bbdba1 02/25/23 10:54:56.332 +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:187 -Dec 14 16:39:06.253: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "disruption-2849" for this suite. 12/14/22 16:39:06.259 -{"msg":"PASSED [sig-apps] DisruptionController should update/patch PodDisruptionBudget status [Conformance]","completed":209,"skipped":3881,"failed":0} +Feb 25 10:54:56.348: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "sched-pred-7417" for this suite. 02/25/23 10:54:56.369 +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:83 +{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance]","completed":207,"skipped":3851,"failed":0} ------------------------------ -• [4.185 seconds] -[sig-apps] DisruptionController -test/e2e/apps/framework.go:23 - should update/patch PodDisruptionBudget status [Conformance] - test/e2e/apps/disruption.go:163 +• [SLOW TEST] [312.703 seconds] +[sig-scheduling] SchedulerPredicates [Serial] +test/e2e/scheduling/framework.go:40 + validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] + test/e2e/scheduling/predicates.go:699 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] DisruptionController + [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:39:02.086 - Dec 14 16:39:02.086: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename disruption 12/14/22 16:39:02.089 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:39:02.11 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:39:02.114 - [BeforeEach] [sig-apps] DisruptionController - test/e2e/apps/disruption.go:71 - [It] should update/patch PodDisruptionBudget status [Conformance] - test/e2e/apps/disruption.go:163 - STEP: Waiting for the pdb to be processed 12/14/22 16:39:02.124 - STEP: Updating PodDisruptionBudget status 12/14/22 16:39:04.14 - STEP: Waiting for all pods to be running 12/14/22 16:39:04.182 - Dec 14 16:39:04.195: INFO: running pods: 0 < 1 - STEP: locating a running pod 12/14/22 16:39:06.203 - STEP: Waiting for the pdb to be processed 12/14/22 16:39:06.221 - STEP: Patching PodDisruptionBudget status 12/14/22 16:39:06.231 - STEP: Waiting for the pdb to be processed 12/14/22 16:39:06.248 - [AfterEach] [sig-apps] DisruptionController + STEP: Creating a kubernetes client 02/25/23 10:49:43.684 + Feb 25 10:49:43.684: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename sched-pred 02/25/23 10:49:43.688 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:49:43.774 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:49:43.781 + [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:92 + Feb 25 10:49:43.786: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready + Feb 25 10:49:43.807: INFO: Waiting for terminating namespaces to be deleted... + Feb 25 10:49:43.816: INFO: + Logging pods the apiserver thinks is on node diek3dua9gay-1 before test + Feb 25 10:49:43.844: INFO: echo-other-node-cd69fcf6b-hsnk9 from cilium-test started at 2023-02-25 09:38:06 +0000 UTC (2 container statuses recorded) + Feb 25 10:49:43.844: INFO: Container dns-test-server ready: true, restart count 0 + Feb 25 10:49:43.844: INFO: Container echo-other-node ready: true, restart count 0 + Feb 25 10:49:43.844: INFO: cilium-node-init-s84z5 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.844: INFO: Container node-init ready: true, restart count 0 + Feb 25 10:49:43.844: INFO: cilium-r4lc4 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.844: INFO: Container cilium-agent ready: true, restart count 0 + Feb 25 10:49:43.844: INFO: coredns-565d847f94-qqvf6 from kube-system started at 2023-02-25 08:49:06 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.844: INFO: Container coredns ready: true, restart count 0 + Feb 25 10:49:43.845: INFO: kube-addon-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.845: INFO: Container kube-addon-manager ready: true, restart count 0 + Feb 25 10:49:43.845: INFO: kube-apiserver-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:09 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.845: INFO: Container kube-apiserver ready: true, restart count 0 + Feb 25 10:49:43.845: INFO: kube-controller-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:49:55 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.845: INFO: Container kube-controller-manager ready: true, restart count 0 + Feb 25 10:49:43.845: INFO: kube-proxy-cm4qg from kube-system started at 2023-02-25 08:46:38 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.845: INFO: Container kube-proxy ready: true, restart count 0 + Feb 25 10:49:43.845: INFO: kube-scheduler-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:27 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.845: INFO: Container kube-scheduler ready: true, restart count 0 + Feb 25 10:49:43.845: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-wznrx from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 10:49:43.846: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 10:49:43.846: INFO: Container systemd-logs ready: true, restart count 0 + Feb 25 10:49:43.846: INFO: + Logging pods the apiserver thinks is on node diek3dua9gay-2 before test + Feb 25 10:49:43.907: INFO: cilium-9tkfg from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.908: INFO: Container cilium-agent ready: true, restart count 0 + Feb 25 10:49:43.908: INFO: cilium-node-init-8b7n2 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.908: INFO: Container node-init ready: true, restart count 0 + Feb 25 10:49:43.908: INFO: cilium-operator-59d7d77cdc-lppbd from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.908: INFO: Container cilium-operator ready: true, restart count 0 + Feb 25 10:49:43.908: INFO: coredns-565d847f94-5nh2n from kube-system started at 2023-02-25 08:48:51 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.908: INFO: Container coredns ready: true, restart count 0 + Feb 25 10:49:43.908: INFO: kube-addon-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.908: INFO: Container kube-addon-manager ready: true, restart count 0 + Feb 25 10:49:43.908: INFO: kube-apiserver-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.908: INFO: Container kube-apiserver ready: true, restart count 0 + Feb 25 10:49:43.908: INFO: kube-controller-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.908: INFO: Container kube-controller-manager ready: true, restart count 0 + Feb 25 10:49:43.908: INFO: kube-proxy-d66m7 from kube-system started at 2023-02-25 08:46:57 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.909: INFO: Container kube-proxy ready: true, restart count 0 + Feb 25 10:49:43.909: INFO: kube-scheduler-diek3dua9gay-2 from kube-system started at 2023-02-25 08:50:15 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.909: INFO: Container kube-scheduler ready: true, restart count 0 + Feb 25 10:49:43.909: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-mml8w from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 10:49:43.909: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 10:49:43.909: INFO: Container systemd-logs ready: true, restart count 0 + Feb 25 10:49:43.909: INFO: + Logging pods the apiserver thinks is on node diek3dua9gay-3 before test + Feb 25 10:49:43.955: INFO: client-7b78db77d5-27xbt from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.955: INFO: Container client ready: true, restart count 0 + Feb 25 10:49:43.955: INFO: client2-78f748dd67-skwz2 from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.955: INFO: Container client2 ready: true, restart count 0 + Feb 25 10:49:43.955: INFO: echo-same-node-85bc9b6b56-wqmfq from cilium-test started at 2023-02-25 09:57:46 +0000 UTC (2 container statuses recorded) + Feb 25 10:49:43.955: INFO: Container dns-test-server ready: true, restart count 0 + Feb 25 10:49:43.955: INFO: Container echo-same-node ready: true, restart count 0 + Feb 25 10:49:43.955: INFO: cilium-2cncz from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.955: INFO: Container cilium-agent ready: true, restart count 0 + Feb 25 10:49:43.955: INFO: cilium-node-init-979mb from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.955: INFO: Container node-init ready: true, restart count 0 + Feb 25 10:49:43.955: INFO: kube-proxy-22s9q from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.955: INFO: Container kube-proxy ready: true, restart count 0 + Feb 25 10:49:43.955: INFO: sonobuoy from sonobuoy started at 2023-02-25 09:45:14 +0000 UTC (1 container statuses recorded) + Feb 25 10:49:43.955: INFO: Container kube-sonobuoy ready: true, restart count 0 + Feb 25 10:49:43.955: INFO: sonobuoy-e2e-job-0b3c30975b5c4880 from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 10:49:43.955: INFO: Container e2e ready: true, restart count 0 + Feb 25 10:49:43.955: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 10:49:43.955: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-9pdjb from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 10:49:43.955: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 10:49:43.955: INFO: Container systemd-logs ready: true, restart count 0 + [It] validates that there exists conflict between pods with same hostPort and protocol but one using 0.0.0.0 hostIP [Conformance] + test/e2e/scheduling/predicates.go:699 + STEP: Trying to launch a pod without a label to get a node which can launch it. 02/25/23 10:49:43.955 + Feb 25 10:49:43.983: INFO: Waiting up to 1m0s for pod "without-label" in namespace "sched-pred-7417" to be "running" + Feb 25 10:49:43.999: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 15.815522ms + Feb 25 10:49:46.015: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031180874s + Feb 25 10:49:48.009: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 4.025307926s + Feb 25 10:49:50.008: INFO: Pod "without-label": Phase="Running", Reason="", readiness=true. Elapsed: 6.024639364s + Feb 25 10:49:50.008: INFO: Pod "without-label" satisfied condition "running" + STEP: Explicitly delete pod here to free the resource it takes. 02/25/23 10:49:50.022 + STEP: Trying to apply a random label on the found node. 02/25/23 10:49:50.069 + STEP: verifying the node has the label kubernetes.io/e2e-cd126814-d2ad-4e65-b03d-da1668bbdba1 95 02/25/23 10:49:50.146 + STEP: Trying to create a pod(pod4) with hostport 54322 and hostIP 0.0.0.0(empty string here) and expect scheduled 02/25/23 10:49:50.194 + Feb 25 10:49:50.219: INFO: Waiting up to 5m0s for pod "pod4" in namespace "sched-pred-7417" to be "not pending" + Feb 25 10:49:50.244: INFO: Pod "pod4": Phase="Pending", Reason="", readiness=false. Elapsed: 24.867341ms + Feb 25 10:49:52.254: INFO: Pod "pod4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.0355768s + Feb 25 10:49:54.255: INFO: Pod "pod4": Phase="Pending", Reason="", readiness=false. Elapsed: 4.035668718s + Feb 25 10:49:56.254: INFO: Pod "pod4": Phase="Running", Reason="", readiness=false. Elapsed: 6.035123734s + Feb 25 10:49:56.254: INFO: Pod "pod4" satisfied condition "not pending" + STEP: Trying to create another pod(pod5) with hostport 54322 but hostIP 192.168.121.44 on the node which pod4 resides and expect not scheduled 02/25/23 10:49:56.254 + Feb 25 10:49:56.267: INFO: Waiting up to 5m0s for pod "pod5" in namespace "sched-pred-7417" to be "not pending" + Feb 25 10:49:56.277: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 10.372375ms + Feb 25 10:49:58.285: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017754162s + Feb 25 10:50:00.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4.022980694s + Feb 25 10:50:02.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 6.018774556s + Feb 25 10:50:04.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 8.019494425s + Feb 25 10:50:06.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 10.023413684s + Feb 25 10:50:08.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 12.021006876s + Feb 25 10:50:10.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 14.021083628s + Feb 25 10:50:12.291: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 16.024000127s + Feb 25 10:50:14.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 18.021954878s + Feb 25 10:50:16.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 20.019869062s + Feb 25 10:50:18.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 22.019178394s + Feb 25 10:50:20.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 24.018806555s + Feb 25 10:50:22.295: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 26.027456039s + Feb 25 10:50:24.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 28.020874362s + Feb 25 10:50:26.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 30.019443817s + Feb 25 10:50:28.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 32.020507907s + Feb 25 10:50:30.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 34.020302952s + Feb 25 10:50:32.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 36.023157663s + Feb 25 10:50:34.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 38.020646692s + Feb 25 10:50:36.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 40.020758481s + Feb 25 10:50:38.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 42.021574844s + Feb 25 10:50:40.296: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 44.028637513s + Feb 25 10:50:42.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 46.021122162s + Feb 25 10:50:44.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 48.01910861s + Feb 25 10:50:46.292: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 50.024741857s + Feb 25 10:50:48.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 52.021488238s + Feb 25 10:50:50.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 54.019052153s + Feb 25 10:50:52.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 56.02222613s + Feb 25 10:50:54.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 58.023262639s + Feb 25 10:50:56.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m0.021449895s + Feb 25 10:50:58.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m2.022250317s + Feb 25 10:51:00.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m4.020838143s + Feb 25 10:51:02.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m6.019857203s + Feb 25 10:51:04.299: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m8.031765945s + Feb 25 10:51:06.285: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m10.01751915s + Feb 25 10:51:08.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m12.021394597s + Feb 25 10:51:10.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m14.01854391s + Feb 25 10:51:12.285: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m16.018260224s + Feb 25 10:51:14.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m18.021428968s + Feb 25 10:51:16.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m20.020308355s + Feb 25 10:51:18.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m22.019923977s + Feb 25 10:51:20.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m24.020542738s + Feb 25 10:51:22.294: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m26.026754742s + Feb 25 10:51:24.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m28.020136091s + Feb 25 10:51:26.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m30.018950333s + Feb 25 10:51:28.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m32.019143054s + Feb 25 10:51:30.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m34.020616866s + Feb 25 10:51:32.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m36.02041168s + Feb 25 10:51:34.300: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m38.032624749s + Feb 25 10:51:36.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m40.020896983s + Feb 25 10:51:38.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m42.0187815s + Feb 25 10:51:40.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m44.019927403s + Feb 25 10:51:42.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m46.021048371s + Feb 25 10:51:44.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m48.02098516s + Feb 25 10:51:46.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m50.021728367s + Feb 25 10:51:48.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m52.01898498s + Feb 25 10:51:50.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m54.019502585s + Feb 25 10:51:52.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m56.018917293s + Feb 25 10:51:54.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 1m58.019156606s + Feb 25 10:51:56.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m0.019696432s + Feb 25 10:51:58.284: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m2.017294273s + Feb 25 10:52:00.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m4.022107645s + Feb 25 10:52:02.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m6.021368374s + Feb 25 10:52:04.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m8.02255758s + Feb 25 10:52:06.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m10.021543676s + Feb 25 10:52:08.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m12.020435477s + Feb 25 10:52:10.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m14.020006886s + Feb 25 10:52:12.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m16.019260452s + Feb 25 10:52:14.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m18.020325224s + Feb 25 10:52:16.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m20.018903392s + Feb 25 10:52:18.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m22.021393285s + Feb 25 10:52:20.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m24.019728092s + Feb 25 10:52:22.285: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m26.018243995s + Feb 25 10:52:24.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m28.019446298s + Feb 25 10:52:26.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m30.019156597s + Feb 25 10:52:28.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m32.019447136s + Feb 25 10:52:30.291: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m34.023914823s + Feb 25 10:52:32.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m36.018574634s + Feb 25 10:52:34.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m38.021058259s + Feb 25 10:52:36.291: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m40.023578732s + Feb 25 10:52:38.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m42.020325603s + Feb 25 10:52:40.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m44.019960249s + Feb 25 10:52:42.292: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m46.025201232s + Feb 25 10:52:44.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m48.020318496s + Feb 25 10:52:46.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m50.019080334s + Feb 25 10:52:48.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m52.018685964s + Feb 25 10:52:50.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m54.021673044s + Feb 25 10:52:52.293: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m56.026317461s + Feb 25 10:52:54.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 2m58.022281364s + Feb 25 10:52:56.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m0.019686908s + Feb 25 10:52:58.291: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m2.023920969s + Feb 25 10:53:00.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m4.01973143s + Feb 25 10:53:02.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m6.020784807s + Feb 25 10:53:04.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m8.021606104s + Feb 25 10:53:06.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m10.018510495s + Feb 25 10:53:08.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m12.019752344s + Feb 25 10:53:10.284: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m14.016683863s + Feb 25 10:53:12.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m16.019354017s + Feb 25 10:53:14.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m18.022577365s + Feb 25 10:53:16.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m20.022692292s + Feb 25 10:53:18.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m22.020882238s + Feb 25 10:53:20.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m24.018833112s + Feb 25 10:53:22.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m26.019092518s + Feb 25 10:53:24.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m28.019855317s + Feb 25 10:53:26.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m30.019760913s + Feb 25 10:53:28.285: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m32.017764967s + Feb 25 10:53:30.285: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m34.017666813s + Feb 25 10:53:32.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m36.019768382s + Feb 25 10:53:34.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m38.018762385s + Feb 25 10:53:36.297: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m40.029841512s + Feb 25 10:53:38.285: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m42.018399347s + Feb 25 10:53:40.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m44.020382494s + Feb 25 10:53:42.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m46.018972935s + Feb 25 10:53:44.323: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m48.055950135s + Feb 25 10:53:46.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m50.021043859s + Feb 25 10:53:48.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m52.020673168s + Feb 25 10:53:50.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m54.019314145s + Feb 25 10:53:52.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m56.019043667s + Feb 25 10:53:54.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 3m58.020658765s + Feb 25 10:53:56.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m0.020677455s + Feb 25 10:53:58.284: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m2.017354685s + Feb 25 10:54:00.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m4.021113242s + Feb 25 10:54:02.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m6.020427842s + Feb 25 10:54:04.292: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m8.024882749s + Feb 25 10:54:06.295: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m10.028206471s + Feb 25 10:54:08.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m12.020606441s + Feb 25 10:54:10.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m14.021653659s + Feb 25 10:54:12.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m16.02263428s + Feb 25 10:54:14.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m18.019703961s + Feb 25 10:54:16.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m20.020184764s + Feb 25 10:54:18.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m22.021723033s + Feb 25 10:54:20.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m24.02017799s + Feb 25 10:54:22.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m26.020964983s + Feb 25 10:54:24.292: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m28.025303638s + Feb 25 10:54:26.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m30.020061323s + Feb 25 10:54:28.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m32.018710104s + Feb 25 10:54:30.291: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m34.023600464s + Feb 25 10:54:32.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m36.019418233s + Feb 25 10:54:34.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m38.019054754s + Feb 25 10:54:36.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m40.018725241s + Feb 25 10:54:38.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m42.020555064s + Feb 25 10:54:40.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m44.019730584s + Feb 25 10:54:42.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m46.020168718s + Feb 25 10:54:44.288: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m48.020910014s + Feb 25 10:54:46.286: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m50.018767923s + Feb 25 10:54:48.291: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m52.024337836s + Feb 25 10:54:50.289: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m54.021607736s + Feb 25 10:54:52.301: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m56.03390871s + Feb 25 10:54:54.290: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 4m58.022824183s + Feb 25 10:54:56.287: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 5m0.019991881s + Feb 25 10:54:56.296: INFO: Pod "pod5": Phase="Pending", Reason="", readiness=false. Elapsed: 5m0.028573083s + STEP: removing the label kubernetes.io/e2e-cd126814-d2ad-4e65-b03d-da1668bbdba1 off the node diek3dua9gay-3 02/25/23 10:54:56.296 + STEP: verifying the node doesn't have the label kubernetes.io/e2e-cd126814-d2ad-4e65-b03d-da1668bbdba1 02/25/23 10:54:56.332 + [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:187 - Dec 14 16:39:06.253: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "disruption-2849" for this suite. 12/14/22 16:39:06.259 + Feb 25 10:54:56.348: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "sched-pred-7417" for this suite. 02/25/23 10:54:56.369 + [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:83 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSS +SSSSSSSSS ------------------------------ [sig-node] Probing container - with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:104 + should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:131 [BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:39:06.272 -Dec 14 16:39:06.272: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-probe 12/14/22 16:39:06.276 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:39:06.314 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:39:06.317 +STEP: Creating a kubernetes client 02/25/23 10:54:56.389 +Feb 25 10:54:56.389: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-probe 02/25/23 10:54:56.419 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:54:56.454 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:54:56.462 [BeforeEach] [sig-node] Probing container test/e2e/common/node/container_probe.go:59 -[It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:104 +[It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:131 +STEP: Creating pod busybox-84a4490e-8857-47fc-ba3e-99788b9797c0 in namespace container-probe-57 02/25/23 10:54:56.473 +Feb 25 10:54:56.494: INFO: Waiting up to 5m0s for pod "busybox-84a4490e-8857-47fc-ba3e-99788b9797c0" in namespace "container-probe-57" to be "not pending" +Feb 25 10:54:56.503: INFO: Pod "busybox-84a4490e-8857-47fc-ba3e-99788b9797c0": Phase="Pending", Reason="", readiness=false. Elapsed: 8.206449ms +Feb 25 10:54:58.513: INFO: Pod "busybox-84a4490e-8857-47fc-ba3e-99788b9797c0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019024622s +Feb 25 10:55:00.512: INFO: Pod "busybox-84a4490e-8857-47fc-ba3e-99788b9797c0": Phase="Running", Reason="", readiness=true. Elapsed: 4.017129006s +Feb 25 10:55:00.512: INFO: Pod "busybox-84a4490e-8857-47fc-ba3e-99788b9797c0" satisfied condition "not pending" +Feb 25 10:55:00.512: INFO: Started pod busybox-84a4490e-8857-47fc-ba3e-99788b9797c0 in namespace container-probe-57 +STEP: checking the pod's current state and verifying that restartCount is present 02/25/23 10:55:00.512 +Feb 25 10:55:00.519: INFO: Initial restart count of pod busybox-84a4490e-8857-47fc-ba3e-99788b9797c0 is 0 +Feb 25 10:55:48.768: INFO: Restart count of pod container-probe-57/busybox-84a4490e-8857-47fc-ba3e-99788b9797c0 is now 1 (48.248933154s elapsed) +STEP: deleting the pod 02/25/23 10:55:48.768 [AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 -Dec 14 16:40:06.346: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-probe-8459" for this suite. 12/14/22 16:40:06.353 -{"msg":"PASSED [sig-node] Probing container with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance]","completed":210,"skipped":3895,"failed":0} +Feb 25 10:55:48.815: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-57" for this suite. 02/25/23 10:55:48.827 +{"msg":"PASSED [sig-node] Probing container should be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","completed":208,"skipped":3860,"failed":0} ------------------------------ -• [SLOW TEST] [60.090 seconds] +• [SLOW TEST] [52.482 seconds] [sig-node] Probing container test/e2e/common/node/framework.go:23 - with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:104 + should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:131 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:39:06.272 - Dec 14 16:39:06.272: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-probe 12/14/22 16:39:06.276 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:39:06.314 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:39:06.317 + STEP: Creating a kubernetes client 02/25/23 10:54:56.389 + Feb 25 10:54:56.389: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-probe 02/25/23 10:54:56.419 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:54:56.454 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:54:56.462 [BeforeEach] [sig-node] Probing container test/e2e/common/node/container_probe.go:59 - [It] with readiness probe that fails should never be ready and never restart [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:104 + [It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:131 + STEP: Creating pod busybox-84a4490e-8857-47fc-ba3e-99788b9797c0 in namespace container-probe-57 02/25/23 10:54:56.473 + Feb 25 10:54:56.494: INFO: Waiting up to 5m0s for pod "busybox-84a4490e-8857-47fc-ba3e-99788b9797c0" in namespace "container-probe-57" to be "not pending" + Feb 25 10:54:56.503: INFO: Pod "busybox-84a4490e-8857-47fc-ba3e-99788b9797c0": Phase="Pending", Reason="", readiness=false. Elapsed: 8.206449ms + Feb 25 10:54:58.513: INFO: Pod "busybox-84a4490e-8857-47fc-ba3e-99788b9797c0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019024622s + Feb 25 10:55:00.512: INFO: Pod "busybox-84a4490e-8857-47fc-ba3e-99788b9797c0": Phase="Running", Reason="", readiness=true. Elapsed: 4.017129006s + Feb 25 10:55:00.512: INFO: Pod "busybox-84a4490e-8857-47fc-ba3e-99788b9797c0" satisfied condition "not pending" + Feb 25 10:55:00.512: INFO: Started pod busybox-84a4490e-8857-47fc-ba3e-99788b9797c0 in namespace container-probe-57 + STEP: checking the pod's current state and verifying that restartCount is present 02/25/23 10:55:00.512 + Feb 25 10:55:00.519: INFO: Initial restart count of pod busybox-84a4490e-8857-47fc-ba3e-99788b9797c0 is 0 + Feb 25 10:55:48.768: INFO: Restart count of pod container-probe-57/busybox-84a4490e-8857-47fc-ba3e-99788b9797c0 is now 1 (48.248933154s elapsed) + STEP: deleting the pod 02/25/23 10:55:48.768 [AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 - Dec 14 16:40:06.346: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-probe-8459" for this suite. 12/14/22 16:40:06.353 + Feb 25 10:55:48.815: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-probe-57" for this suite. 02/25/23 10:55:48.827 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSS +S ------------------------------ -[sig-node] Downward API - should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:216 -[BeforeEach] [sig-node] Downward API +[sig-apps] ReplicaSet + should adopt matching pods on creation and release no longer matching pods [Conformance] + test/e2e/apps/replica_set.go:131 +[BeforeEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:40:06.366 -Dec 14 16:40:06.366: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 16:40:06.371 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:40:06.401 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:40:06.404 -[It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:216 -STEP: Creating a pod to test downward api env vars 12/14/22 16:40:06.408 -Dec 14 16:40:06.427: INFO: Waiting up to 5m0s for pod "downward-api-6e474f21-2df9-420c-a8e4-776468ef4168" in namespace "downward-api-9534" to be "Succeeded or Failed" -Dec 14 16:40:06.432: INFO: Pod "downward-api-6e474f21-2df9-420c-a8e4-776468ef4168": Phase="Pending", Reason="", readiness=false. Elapsed: 5.25375ms -Dec 14 16:40:08.438: INFO: Pod "downward-api-6e474f21-2df9-420c-a8e4-776468ef4168": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011370245s -Dec 14 16:40:10.439: INFO: Pod "downward-api-6e474f21-2df9-420c-a8e4-776468ef4168": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012828522s -STEP: Saw pod success 12/14/22 16:40:10.439 -Dec 14 16:40:10.440: INFO: Pod "downward-api-6e474f21-2df9-420c-a8e4-776468ef4168" satisfied condition "Succeeded or Failed" -Dec 14 16:40:10.445: INFO: Trying to get logs from node kiedai4iakae-3 pod downward-api-6e474f21-2df9-420c-a8e4-776468ef4168 container dapi-container: -STEP: delete the pod 12/14/22 16:40:10.47 -Dec 14 16:40:10.489: INFO: Waiting for pod downward-api-6e474f21-2df9-420c-a8e4-776468ef4168 to disappear -Dec 14 16:40:10.496: INFO: Pod downward-api-6e474f21-2df9-420c-a8e4-776468ef4168 no longer exists -[AfterEach] [sig-node] Downward API +STEP: Creating a kubernetes client 02/25/23 10:55:48.872 +Feb 25 10:55:48.873: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename replicaset 02/25/23 10:55:48.876 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:55:49.031 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:55:49.038 +[It] should adopt matching pods on creation and release no longer matching pods [Conformance] + test/e2e/apps/replica_set.go:131 +STEP: Given a Pod with a 'name' label pod-adoption-release is created 02/25/23 10:55:49.074 +Feb 25 10:55:49.128: INFO: Waiting up to 5m0s for pod "pod-adoption-release" in namespace "replicaset-8573" to be "running and ready" +Feb 25 10:55:49.138: INFO: Pod "pod-adoption-release": Phase="Pending", Reason="", readiness=false. Elapsed: 10.014427ms +Feb 25 10:55:49.138: INFO: The phase of Pod pod-adoption-release is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:55:51.146: INFO: Pod "pod-adoption-release": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018281275s +Feb 25 10:55:51.146: INFO: The phase of Pod pod-adoption-release is Pending, waiting for it to be Running (with Ready = true) +Feb 25 10:55:53.259: INFO: Pod "pod-adoption-release": Phase="Running", Reason="", readiness=true. Elapsed: 4.13175192s +Feb 25 10:55:53.259: INFO: The phase of Pod pod-adoption-release is Running (Ready = true) +Feb 25 10:55:53.259: INFO: Pod "pod-adoption-release" satisfied condition "running and ready" +STEP: When a replicaset with a matching selector is created 02/25/23 10:55:53.312 +STEP: Then the orphan pod is adopted 02/25/23 10:55:53.379 +STEP: When the matched label of one of its pods change 02/25/23 10:55:54.413 +Feb 25 10:55:54.420: INFO: Pod name pod-adoption-release: Found 1 pods out of 1 +STEP: Then the pod is released 02/25/23 10:55:54.48 +[AfterEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:187 -Dec 14 16:40:10.496: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-9534" for this suite. 12/14/22 16:40:10.507 -{"msg":"PASSED [sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]","completed":211,"skipped":3909,"failed":0} +Feb 25 10:55:55.502: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "replicaset-8573" for this suite. 02/25/23 10:55:55.515 +{"msg":"PASSED [sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance]","completed":209,"skipped":3861,"failed":0} ------------------------------ -• [4.150 seconds] -[sig-node] Downward API -test/e2e/common/node/framework.go:23 - should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:216 +• [SLOW TEST] [6.663 seconds] +[sig-apps] ReplicaSet +test/e2e/apps/framework.go:23 + should adopt matching pods on creation and release no longer matching pods [Conformance] + test/e2e/apps/replica_set.go:131 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Downward API + [BeforeEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:40:06.366 - Dec 14 16:40:06.366: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 16:40:06.371 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:40:06.401 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:40:06.404 - [It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:216 - STEP: Creating a pod to test downward api env vars 12/14/22 16:40:06.408 - Dec 14 16:40:06.427: INFO: Waiting up to 5m0s for pod "downward-api-6e474f21-2df9-420c-a8e4-776468ef4168" in namespace "downward-api-9534" to be "Succeeded or Failed" - Dec 14 16:40:06.432: INFO: Pod "downward-api-6e474f21-2df9-420c-a8e4-776468ef4168": Phase="Pending", Reason="", readiness=false. Elapsed: 5.25375ms - Dec 14 16:40:08.438: INFO: Pod "downward-api-6e474f21-2df9-420c-a8e4-776468ef4168": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011370245s - Dec 14 16:40:10.439: INFO: Pod "downward-api-6e474f21-2df9-420c-a8e4-776468ef4168": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012828522s - STEP: Saw pod success 12/14/22 16:40:10.439 - Dec 14 16:40:10.440: INFO: Pod "downward-api-6e474f21-2df9-420c-a8e4-776468ef4168" satisfied condition "Succeeded or Failed" - Dec 14 16:40:10.445: INFO: Trying to get logs from node kiedai4iakae-3 pod downward-api-6e474f21-2df9-420c-a8e4-776468ef4168 container dapi-container: - STEP: delete the pod 12/14/22 16:40:10.47 - Dec 14 16:40:10.489: INFO: Waiting for pod downward-api-6e474f21-2df9-420c-a8e4-776468ef4168 to disappear - Dec 14 16:40:10.496: INFO: Pod downward-api-6e474f21-2df9-420c-a8e4-776468ef4168 no longer exists - [AfterEach] [sig-node] Downward API + STEP: Creating a kubernetes client 02/25/23 10:55:48.872 + Feb 25 10:55:48.873: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename replicaset 02/25/23 10:55:48.876 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:55:49.031 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:55:49.038 + [It] should adopt matching pods on creation and release no longer matching pods [Conformance] + test/e2e/apps/replica_set.go:131 + STEP: Given a Pod with a 'name' label pod-adoption-release is created 02/25/23 10:55:49.074 + Feb 25 10:55:49.128: INFO: Waiting up to 5m0s for pod "pod-adoption-release" in namespace "replicaset-8573" to be "running and ready" + Feb 25 10:55:49.138: INFO: Pod "pod-adoption-release": Phase="Pending", Reason="", readiness=false. Elapsed: 10.014427ms + Feb 25 10:55:49.138: INFO: The phase of Pod pod-adoption-release is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:55:51.146: INFO: Pod "pod-adoption-release": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018281275s + Feb 25 10:55:51.146: INFO: The phase of Pod pod-adoption-release is Pending, waiting for it to be Running (with Ready = true) + Feb 25 10:55:53.259: INFO: Pod "pod-adoption-release": Phase="Running", Reason="", readiness=true. Elapsed: 4.13175192s + Feb 25 10:55:53.259: INFO: The phase of Pod pod-adoption-release is Running (Ready = true) + Feb 25 10:55:53.259: INFO: Pod "pod-adoption-release" satisfied condition "running and ready" + STEP: When a replicaset with a matching selector is created 02/25/23 10:55:53.312 + STEP: Then the orphan pod is adopted 02/25/23 10:55:53.379 + STEP: When the matched label of one of its pods change 02/25/23 10:55:54.413 + Feb 25 10:55:54.420: INFO: Pod name pod-adoption-release: Found 1 pods out of 1 + STEP: Then the pod is released 02/25/23 10:55:54.48 + [AfterEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:187 - Dec 14 16:40:10.496: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-9534" for this suite. 12/14/22 16:40:10.507 + Feb 25 10:55:55.502: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "replicaset-8573" for this suite. 02/25/23 10:55:55.515 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSS ------------------------------ -[sig-network] Services - should delete a collection of services [Conformance] - test/e2e/network/service.go:3641 -[BeforeEach] [sig-network] Services +[sig-storage] EmptyDir volumes + should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:96 +[BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:40:10.521 -Dec 14 16:40:10.522: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 16:40:10.523 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:40:10.548 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:40:10.551 -[BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 -[It] should delete a collection of services [Conformance] - test/e2e/network/service.go:3641 -STEP: creating a collection of services 12/14/22 16:40:10.554 -Dec 14 16:40:10.554: INFO: Creating e2e-svc-a-kw9tx -Dec 14 16:40:10.570: INFO: Creating e2e-svc-b-j42lq -Dec 14 16:40:10.607: INFO: Creating e2e-svc-c-ld26f -STEP: deleting service collection 12/14/22 16:40:10.63 -Dec 14 16:40:10.697: INFO: Collection of services has been deleted -[AfterEach] [sig-network] Services +STEP: Creating a kubernetes client 02/25/23 10:55:55.542 +Feb 25 10:55:55.543: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir 02/25/23 10:55:55.548 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:55:55.59 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:55:55.599 +[It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:96 +STEP: Creating a pod to test emptydir 0644 on tmpfs 02/25/23 10:55:55.605 +Feb 25 10:55:55.635: INFO: Waiting up to 5m0s for pod "pod-b42bb6fc-e894-474e-87e9-b9c642f6771a" in namespace "emptydir-9325" to be "Succeeded or Failed" +Feb 25 10:55:55.648: INFO: Pod "pod-b42bb6fc-e894-474e-87e9-b9c642f6771a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.032343ms +Feb 25 10:55:57.661: INFO: Pod "pod-b42bb6fc-e894-474e-87e9-b9c642f6771a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025585224s +Feb 25 10:55:59.679: INFO: Pod "pod-b42bb6fc-e894-474e-87e9-b9c642f6771a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.044480858s +Feb 25 10:56:01.675: INFO: Pod "pod-b42bb6fc-e894-474e-87e9-b9c642f6771a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.040096763s +STEP: Saw pod success 02/25/23 10:56:01.675 +Feb 25 10:56:01.676: INFO: Pod "pod-b42bb6fc-e894-474e-87e9-b9c642f6771a" satisfied condition "Succeeded or Failed" +Feb 25 10:56:01.688: INFO: Trying to get logs from node diek3dua9gay-2 pod pod-b42bb6fc-e894-474e-87e9-b9c642f6771a container test-container: +STEP: delete the pod 02/25/23 10:56:01.739 +Feb 25 10:56:01.787: INFO: Waiting for pod pod-b42bb6fc-e894-474e-87e9-b9c642f6771a to disappear +Feb 25 10:56:01.814: INFO: Pod pod-b42bb6fc-e894-474e-87e9-b9c642f6771a no longer exists +[AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 -Dec 14 16:40:10.697: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-9708" for this suite. 12/14/22 16:40:10.706 -[AfterEach] [sig-network] Services - test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should delete a collection of services [Conformance]","completed":212,"skipped":3950,"failed":0} +Feb 25 10:56:01.815: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-9325" for this suite. 02/25/23 10:56:01.837 +{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","completed":210,"skipped":3871,"failed":0} ------------------------------ -• [0.198 seconds] -[sig-network] Services -test/e2e/network/common/framework.go:23 - should delete a collection of services [Conformance] - test/e2e/network/service.go:3641 +• [SLOW TEST] [6.311 seconds] +[sig-storage] EmptyDir volumes +test/e2e/common/storage/framework.go:23 + should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:96 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Services + [BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:40:10.521 - Dec 14 16:40:10.522: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 16:40:10.523 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:40:10.548 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:40:10.551 - [BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 - [It] should delete a collection of services [Conformance] - test/e2e/network/service.go:3641 - STEP: creating a collection of services 12/14/22 16:40:10.554 - Dec 14 16:40:10.554: INFO: Creating e2e-svc-a-kw9tx - Dec 14 16:40:10.570: INFO: Creating e2e-svc-b-j42lq - Dec 14 16:40:10.607: INFO: Creating e2e-svc-c-ld26f - STEP: deleting service collection 12/14/22 16:40:10.63 - Dec 14 16:40:10.697: INFO: Collection of services has been deleted - [AfterEach] [sig-network] Services + STEP: Creating a kubernetes client 02/25/23 10:55:55.542 + Feb 25 10:55:55.543: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir 02/25/23 10:55:55.548 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:55:55.59 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:55:55.599 + [It] should support (root,0644,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:96 + STEP: Creating a pod to test emptydir 0644 on tmpfs 02/25/23 10:55:55.605 + Feb 25 10:55:55.635: INFO: Waiting up to 5m0s for pod "pod-b42bb6fc-e894-474e-87e9-b9c642f6771a" in namespace "emptydir-9325" to be "Succeeded or Failed" + Feb 25 10:55:55.648: INFO: Pod "pod-b42bb6fc-e894-474e-87e9-b9c642f6771a": Phase="Pending", Reason="", readiness=false. Elapsed: 13.032343ms + Feb 25 10:55:57.661: INFO: Pod "pod-b42bb6fc-e894-474e-87e9-b9c642f6771a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.025585224s + Feb 25 10:55:59.679: INFO: Pod "pod-b42bb6fc-e894-474e-87e9-b9c642f6771a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.044480858s + Feb 25 10:56:01.675: INFO: Pod "pod-b42bb6fc-e894-474e-87e9-b9c642f6771a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.040096763s + STEP: Saw pod success 02/25/23 10:56:01.675 + Feb 25 10:56:01.676: INFO: Pod "pod-b42bb6fc-e894-474e-87e9-b9c642f6771a" satisfied condition "Succeeded or Failed" + Feb 25 10:56:01.688: INFO: Trying to get logs from node diek3dua9gay-2 pod pod-b42bb6fc-e894-474e-87e9-b9c642f6771a container test-container: + STEP: delete the pod 02/25/23 10:56:01.739 + Feb 25 10:56:01.787: INFO: Waiting for pod pod-b42bb6fc-e894-474e-87e9-b9c642f6771a to disappear + Feb 25 10:56:01.814: INFO: Pod pod-b42bb6fc-e894-474e-87e9-b9c642f6771a no longer exists + [AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 - Dec 14 16:40:10.697: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-9708" for this suite. 12/14/22 16:40:10.706 - [AfterEach] [sig-network] Services - test/e2e/network/service.go:762 + Feb 25 10:56:01.815: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-9325" for this suite. 02/25/23 10:56:01.837 << End Captured GinkgoWriter Output ------------------------------ -SSSSS +SSSSSSSSSSSSSSSSSSS ------------------------------ [sig-node] RuntimeClass - should support RuntimeClasses API operations [Conformance] - test/e2e/common/node/runtimeclass.go:189 + should reject a Pod requesting a deleted RuntimeClass [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:156 [BeforeEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:40:10.724 -Dec 14 16:40:10.724: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename runtimeclass 12/14/22 16:40:10.727 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:40:10.749 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:40:10.752 -[It] should support RuntimeClasses API operations [Conformance] - test/e2e/common/node/runtimeclass.go:189 -STEP: getting /apis 12/14/22 16:40:10.756 -STEP: getting /apis/node.k8s.io 12/14/22 16:40:10.758 -STEP: getting /apis/node.k8s.io/v1 12/14/22 16:40:10.76 -STEP: creating 12/14/22 16:40:10.761 -STEP: watching 12/14/22 16:40:10.78 -Dec 14 16:40:10.781: INFO: starting watch -STEP: getting 12/14/22 16:40:10.788 -STEP: listing 12/14/22 16:40:10.793 -STEP: patching 12/14/22 16:40:10.798 -STEP: updating 12/14/22 16:40:10.805 -Dec 14 16:40:10.813: INFO: waiting for watch events with expected annotations -STEP: deleting 12/14/22 16:40:10.813 -STEP: deleting a collection 12/14/22 16:40:10.832 +STEP: Creating a kubernetes client 02/25/23 10:56:01.862 +Feb 25 10:56:01.862: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename runtimeclass 02/25/23 10:56:01.864 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:56:01.908 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:56:01.912 +[It] should reject a Pod requesting a deleted RuntimeClass [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:156 +STEP: Deleting RuntimeClass runtimeclass-3953-delete-me 02/25/23 10:56:01.925 +STEP: Waiting for the RuntimeClass to disappear 02/25/23 10:56:01.947 [AfterEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:187 -Dec 14 16:40:10.857: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "runtimeclass-1978" for this suite. 12/14/22 16:40:10.864 -{"msg":"PASSED [sig-node] RuntimeClass should support RuntimeClasses API operations [Conformance]","completed":213,"skipped":3955,"failed":0} +Feb 25 10:56:02.023: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "runtimeclass-3953" for this suite. 02/25/23 10:56:02.039 +{"msg":"PASSED [sig-node] RuntimeClass should reject a Pod requesting a deleted RuntimeClass [NodeConformance] [Conformance]","completed":211,"skipped":3890,"failed":0} ------------------------------ -• [0.151 seconds] +• [0.197 seconds] [sig-node] RuntimeClass test/e2e/common/node/framework.go:23 - should support RuntimeClasses API operations [Conformance] - test/e2e/common/node/runtimeclass.go:189 + should reject a Pod requesting a deleted RuntimeClass [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:156 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:40:10.724 - Dec 14 16:40:10.724: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename runtimeclass 12/14/22 16:40:10.727 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:40:10.749 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:40:10.752 - [It] should support RuntimeClasses API operations [Conformance] - test/e2e/common/node/runtimeclass.go:189 - STEP: getting /apis 12/14/22 16:40:10.756 - STEP: getting /apis/node.k8s.io 12/14/22 16:40:10.758 - STEP: getting /apis/node.k8s.io/v1 12/14/22 16:40:10.76 - STEP: creating 12/14/22 16:40:10.761 - STEP: watching 12/14/22 16:40:10.78 - Dec 14 16:40:10.781: INFO: starting watch - STEP: getting 12/14/22 16:40:10.788 - STEP: listing 12/14/22 16:40:10.793 - STEP: patching 12/14/22 16:40:10.798 - STEP: updating 12/14/22 16:40:10.805 - Dec 14 16:40:10.813: INFO: waiting for watch events with expected annotations - STEP: deleting 12/14/22 16:40:10.813 - STEP: deleting a collection 12/14/22 16:40:10.832 + STEP: Creating a kubernetes client 02/25/23 10:56:01.862 + Feb 25 10:56:01.862: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename runtimeclass 02/25/23 10:56:01.864 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:56:01.908 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:56:01.912 + [It] should reject a Pod requesting a deleted RuntimeClass [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:156 + STEP: Deleting RuntimeClass runtimeclass-3953-delete-me 02/25/23 10:56:01.925 + STEP: Waiting for the RuntimeClass to disappear 02/25/23 10:56:01.947 [AfterEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:187 - Dec 14 16:40:10.857: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "runtimeclass-1978" for this suite. 12/14/22 16:40:10.864 - << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSSSSSSSSSS ------------------------------- -[sig-storage] EmptyDir volumes - volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:86 -[BeforeEach] [sig-storage] EmptyDir volumes - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:40:10.875 -Dec 14 16:40:10.875: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir 12/14/22 16:40:10.878 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:40:10.9 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:40:10.903 -[It] volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:86 -STEP: Creating a pod to test emptydir volume type on tmpfs 12/14/22 16:40:10.907 -Dec 14 16:40:10.919: INFO: Waiting up to 5m0s for pod "pod-4b2a8370-5029-46f5-8051-fbac6543dd96" in namespace "emptydir-1361" to be "Succeeded or Failed" -Dec 14 16:40:10.922: INFO: Pod "pod-4b2a8370-5029-46f5-8051-fbac6543dd96": Phase="Pending", Reason="", readiness=false. Elapsed: 3.821927ms -Dec 14 16:40:12.929: INFO: Pod "pod-4b2a8370-5029-46f5-8051-fbac6543dd96": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01077222s -Dec 14 16:40:14.929: INFO: Pod "pod-4b2a8370-5029-46f5-8051-fbac6543dd96": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010702777s -STEP: Saw pod success 12/14/22 16:40:14.929 -Dec 14 16:40:14.930: INFO: Pod "pod-4b2a8370-5029-46f5-8051-fbac6543dd96" satisfied condition "Succeeded or Failed" -Dec 14 16:40:14.936: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-4b2a8370-5029-46f5-8051-fbac6543dd96 container test-container: -STEP: delete the pod 12/14/22 16:40:14.945 -Dec 14 16:40:14.958: INFO: Waiting for pod pod-4b2a8370-5029-46f5-8051-fbac6543dd96 to disappear -Dec 14 16:40:14.961: INFO: Pod pod-4b2a8370-5029-46f5-8051-fbac6543dd96 no longer exists -[AfterEach] [sig-storage] EmptyDir volumes - test/e2e/framework/framework.go:187 -Dec 14 16:40:14.961: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-1361" for this suite. 12/14/22 16:40:14.966 -{"msg":"PASSED [sig-storage] EmptyDir volumes volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","completed":214,"skipped":3971,"failed":0} ------------------------------- -• [4.101 seconds] -[sig-storage] EmptyDir volumes -test/e2e/common/storage/framework.go:23 - volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:86 - - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir volumes - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:40:10.875 - Dec 14 16:40:10.875: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir 12/14/22 16:40:10.878 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:40:10.9 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:40:10.903 - [It] volume on tmpfs should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:86 - STEP: Creating a pod to test emptydir volume type on tmpfs 12/14/22 16:40:10.907 - Dec 14 16:40:10.919: INFO: Waiting up to 5m0s for pod "pod-4b2a8370-5029-46f5-8051-fbac6543dd96" in namespace "emptydir-1361" to be "Succeeded or Failed" - Dec 14 16:40:10.922: INFO: Pod "pod-4b2a8370-5029-46f5-8051-fbac6543dd96": Phase="Pending", Reason="", readiness=false. Elapsed: 3.821927ms - Dec 14 16:40:12.929: INFO: Pod "pod-4b2a8370-5029-46f5-8051-fbac6543dd96": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01077222s - Dec 14 16:40:14.929: INFO: Pod "pod-4b2a8370-5029-46f5-8051-fbac6543dd96": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010702777s - STEP: Saw pod success 12/14/22 16:40:14.929 - Dec 14 16:40:14.930: INFO: Pod "pod-4b2a8370-5029-46f5-8051-fbac6543dd96" satisfied condition "Succeeded or Failed" - Dec 14 16:40:14.936: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-4b2a8370-5029-46f5-8051-fbac6543dd96 container test-container: - STEP: delete the pod 12/14/22 16:40:14.945 - Dec 14 16:40:14.958: INFO: Waiting for pod pod-4b2a8370-5029-46f5-8051-fbac6543dd96 to disappear - Dec 14 16:40:14.961: INFO: Pod pod-4b2a8370-5029-46f5-8051-fbac6543dd96 no longer exists - [AfterEach] [sig-storage] EmptyDir volumes - test/e2e/framework/framework.go:187 - Dec 14 16:40:14.961: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-1361" for this suite. 12/14/22 16:40:14.966 + Feb 25 10:56:02.023: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "runtimeclass-3953" for this suite. 02/25/23 10:56:02.039 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] Watchers - should be able to restart watching from the last resource version observed by the previous watch [Conformance] - test/e2e/apimachinery/watch.go:191 -[BeforeEach] [sig-api-machinery] Watchers +[sig-scheduling] SchedulerPreemption [Serial] + validates lower priority pod preemption by critical pod [Conformance] + test/e2e/scheduling/preemption.go:218 +[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:40:14.977 -Dec 14 16:40:14.977: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename watch 12/14/22 16:40:14.98 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:40:15.007 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:40:15.01 -[It] should be able to restart watching from the last resource version observed by the previous watch [Conformance] - test/e2e/apimachinery/watch.go:191 -STEP: creating a watch on configmaps 12/14/22 16:40:15.014 -STEP: creating a new configmap 12/14/22 16:40:15.015 -STEP: modifying the configmap once 12/14/22 16:40:15.027 -STEP: closing the watch once it receives two notifications 12/14/22 16:40:15.039 -Dec 14 16:40:15.039: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-7972 2942390c-f9ae-48f9-ba14-cb0af8730fab 22326 0 2022-12-14 16:40:14 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2022-12-14 16:40:14 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} -Dec 14 16:40:15.040: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-7972 2942390c-f9ae-48f9-ba14-cb0af8730fab 22327 0 2022-12-14 16:40:14 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2022-12-14 16:40:14 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} -STEP: modifying the configmap a second time, while the watch is closed 12/14/22 16:40:15.04 -STEP: creating a new watch on configmaps from the last resource version observed by the first watch 12/14/22 16:40:15.05 -STEP: deleting the configmap 12/14/22 16:40:15.051 -STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed 12/14/22 16:40:15.058 -Dec 14 16:40:15.058: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-7972 2942390c-f9ae-48f9-ba14-cb0af8730fab 22328 0 2022-12-14 16:40:14 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2022-12-14 16:40:14 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} -Dec 14 16:40:15.058: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-7972 2942390c-f9ae-48f9-ba14-cb0af8730fab 22329 0 2022-12-14 16:40:14 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2022-12-14 16:40:14 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} -[AfterEach] [sig-api-machinery] Watchers +STEP: Creating a kubernetes client 02/25/23 10:56:02.083 +Feb 25 10:56:02.083: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename sched-preemption 02/25/23 10:56:02.086 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:56:02.127 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:56:02.133 +[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:92 +Feb 25 10:56:02.178: INFO: Waiting up to 1m0s for all nodes to be ready +Feb 25 10:57:02.270: INFO: Waiting for terminating namespaces to be deleted... +[It] validates lower priority pod preemption by critical pod [Conformance] + test/e2e/scheduling/preemption.go:218 +STEP: Create pods that use 4/5 of node resources. 02/25/23 10:57:02.278 +Feb 25 10:57:02.322: INFO: Created pod: pod0-0-sched-preemption-low-priority +Feb 25 10:57:02.346: INFO: Created pod: pod0-1-sched-preemption-medium-priority +Feb 25 10:57:02.395: INFO: Created pod: pod1-0-sched-preemption-medium-priority +Feb 25 10:57:02.416: INFO: Created pod: pod1-1-sched-preemption-medium-priority +Feb 25 10:57:02.459: INFO: Created pod: pod2-0-sched-preemption-medium-priority +Feb 25 10:57:02.486: INFO: Created pod: pod2-1-sched-preemption-medium-priority +STEP: Wait for pods to be scheduled. 02/25/23 10:57:02.487 +Feb 25 10:57:02.487: INFO: Waiting up to 5m0s for pod "pod0-0-sched-preemption-low-priority" in namespace "sched-preemption-3418" to be "running" +Feb 25 10:57:02.507: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 19.623072ms +Feb 25 10:57:04.517: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029645244s +Feb 25 10:57:06.518: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 4.03128666s +Feb 25 10:57:08.515: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 6.028151178s +Feb 25 10:57:10.518: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Running", Reason="", readiness=true. Elapsed: 8.031222122s +Feb 25 10:57:10.518: INFO: Pod "pod0-0-sched-preemption-low-priority" satisfied condition "running" +Feb 25 10:57:10.519: INFO: Waiting up to 5m0s for pod "pod0-1-sched-preemption-medium-priority" in namespace "sched-preemption-3418" to be "running" +Feb 25 10:57:10.528: INFO: Pod "pod0-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 9.494174ms +Feb 25 10:57:10.528: INFO: Pod "pod0-1-sched-preemption-medium-priority" satisfied condition "running" +Feb 25 10:57:10.528: INFO: Waiting up to 5m0s for pod "pod1-0-sched-preemption-medium-priority" in namespace "sched-preemption-3418" to be "running" +Feb 25 10:57:10.538: INFO: Pod "pod1-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 9.277832ms +Feb 25 10:57:10.538: INFO: Pod "pod1-0-sched-preemption-medium-priority" satisfied condition "running" +Feb 25 10:57:10.538: INFO: Waiting up to 5m0s for pod "pod1-1-sched-preemption-medium-priority" in namespace "sched-preemption-3418" to be "running" +Feb 25 10:57:10.547: INFO: Pod "pod1-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 8.79257ms +Feb 25 10:57:10.547: INFO: Pod "pod1-1-sched-preemption-medium-priority" satisfied condition "running" +Feb 25 10:57:10.547: INFO: Waiting up to 5m0s for pod "pod2-0-sched-preemption-medium-priority" in namespace "sched-preemption-3418" to be "running" +Feb 25 10:57:10.554: INFO: Pod "pod2-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 7.76712ms +Feb 25 10:57:10.555: INFO: Pod "pod2-0-sched-preemption-medium-priority" satisfied condition "running" +Feb 25 10:57:10.555: INFO: Waiting up to 5m0s for pod "pod2-1-sched-preemption-medium-priority" in namespace "sched-preemption-3418" to be "running" +Feb 25 10:57:10.562: INFO: Pod "pod2-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 7.036255ms +Feb 25 10:57:10.562: INFO: Pod "pod2-1-sched-preemption-medium-priority" satisfied condition "running" +STEP: Run a critical pod that use same resources as that of a lower priority pod 02/25/23 10:57:10.562 +Feb 25 10:57:10.583: INFO: Waiting up to 2m0s for pod "critical-pod" in namespace "kube-system" to be "running" +Feb 25 10:57:10.596: INFO: Pod "critical-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 12.657902ms +Feb 25 10:57:12.606: INFO: Pod "critical-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022683344s +Feb 25 10:57:14.608: INFO: Pod "critical-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 4.025107863s +Feb 25 10:57:16.606: INFO: Pod "critical-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 6.022833053s +Feb 25 10:57:18.607: INFO: Pod "critical-pod": Phase="Running", Reason="", readiness=true. Elapsed: 8.023858604s +Feb 25 10:57:18.607: INFO: Pod "critical-pod" satisfied condition "running" +[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:187 -Dec 14 16:40:15.059: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "watch-7972" for this suite. 12/14/22 16:40:15.067 -{"msg":"PASSED [sig-api-machinery] Watchers should be able to restart watching from the last resource version observed by the previous watch [Conformance]","completed":215,"skipped":3974,"failed":0} +Feb 25 10:57:18.698: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "sched-preemption-3418" for this suite. 02/25/23 10:57:18.717 +[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:80 +{"msg":"PASSED [sig-scheduling] SchedulerPreemption [Serial] validates lower priority pod preemption by critical pod [Conformance]","completed":212,"skipped":3921,"failed":0} ------------------------------ -• [0.098 seconds] -[sig-api-machinery] Watchers -test/e2e/apimachinery/framework.go:23 - should be able to restart watching from the last resource version observed by the previous watch [Conformance] - test/e2e/apimachinery/watch.go:191 +• [SLOW TEST] [76.795 seconds] +[sig-scheduling] SchedulerPreemption [Serial] +test/e2e/scheduling/framework.go:40 + validates lower priority pod preemption by critical pod [Conformance] + test/e2e/scheduling/preemption.go:218 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Watchers + [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:40:14.977 - Dec 14 16:40:14.977: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename watch 12/14/22 16:40:14.98 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:40:15.007 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:40:15.01 - [It] should be able to restart watching from the last resource version observed by the previous watch [Conformance] - test/e2e/apimachinery/watch.go:191 - STEP: creating a watch on configmaps 12/14/22 16:40:15.014 - STEP: creating a new configmap 12/14/22 16:40:15.015 - STEP: modifying the configmap once 12/14/22 16:40:15.027 - STEP: closing the watch once it receives two notifications 12/14/22 16:40:15.039 - Dec 14 16:40:15.039: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-7972 2942390c-f9ae-48f9-ba14-cb0af8730fab 22326 0 2022-12-14 16:40:14 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2022-12-14 16:40:14 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} - Dec 14 16:40:15.040: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-7972 2942390c-f9ae-48f9-ba14-cb0af8730fab 22327 0 2022-12-14 16:40:14 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2022-12-14 16:40:14 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} - STEP: modifying the configmap a second time, while the watch is closed 12/14/22 16:40:15.04 - STEP: creating a new watch on configmaps from the last resource version observed by the first watch 12/14/22 16:40:15.05 - STEP: deleting the configmap 12/14/22 16:40:15.051 - STEP: Expecting to observe notifications for all changes to the configmap since the first watch closed 12/14/22 16:40:15.058 - Dec 14 16:40:15.058: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-7972 2942390c-f9ae-48f9-ba14-cb0af8730fab 22328 0 2022-12-14 16:40:14 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2022-12-14 16:40:14 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} - Dec 14 16:40:15.058: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-watch-closed watch-7972 2942390c-f9ae-48f9-ba14-cb0af8730fab 22329 0 2022-12-14 16:40:14 +0000 UTC map[watch-this-configmap:watch-closed-and-restarted] map[] [] [] [{e2e.test Update v1 2022-12-14 16:40:14 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} - [AfterEach] [sig-api-machinery] Watchers + STEP: Creating a kubernetes client 02/25/23 10:56:02.083 + Feb 25 10:56:02.083: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename sched-preemption 02/25/23 10:56:02.086 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:56:02.127 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:56:02.133 + [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:92 + Feb 25 10:56:02.178: INFO: Waiting up to 1m0s for all nodes to be ready + Feb 25 10:57:02.270: INFO: Waiting for terminating namespaces to be deleted... + [It] validates lower priority pod preemption by critical pod [Conformance] + test/e2e/scheduling/preemption.go:218 + STEP: Create pods that use 4/5 of node resources. 02/25/23 10:57:02.278 + Feb 25 10:57:02.322: INFO: Created pod: pod0-0-sched-preemption-low-priority + Feb 25 10:57:02.346: INFO: Created pod: pod0-1-sched-preemption-medium-priority + Feb 25 10:57:02.395: INFO: Created pod: pod1-0-sched-preemption-medium-priority + Feb 25 10:57:02.416: INFO: Created pod: pod1-1-sched-preemption-medium-priority + Feb 25 10:57:02.459: INFO: Created pod: pod2-0-sched-preemption-medium-priority + Feb 25 10:57:02.486: INFO: Created pod: pod2-1-sched-preemption-medium-priority + STEP: Wait for pods to be scheduled. 02/25/23 10:57:02.487 + Feb 25 10:57:02.487: INFO: Waiting up to 5m0s for pod "pod0-0-sched-preemption-low-priority" in namespace "sched-preemption-3418" to be "running" + Feb 25 10:57:02.507: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 19.623072ms + Feb 25 10:57:04.517: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 2.029645244s + Feb 25 10:57:06.518: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 4.03128666s + Feb 25 10:57:08.515: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 6.028151178s + Feb 25 10:57:10.518: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Running", Reason="", readiness=true. Elapsed: 8.031222122s + Feb 25 10:57:10.518: INFO: Pod "pod0-0-sched-preemption-low-priority" satisfied condition "running" + Feb 25 10:57:10.519: INFO: Waiting up to 5m0s for pod "pod0-1-sched-preemption-medium-priority" in namespace "sched-preemption-3418" to be "running" + Feb 25 10:57:10.528: INFO: Pod "pod0-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 9.494174ms + Feb 25 10:57:10.528: INFO: Pod "pod0-1-sched-preemption-medium-priority" satisfied condition "running" + Feb 25 10:57:10.528: INFO: Waiting up to 5m0s for pod "pod1-0-sched-preemption-medium-priority" in namespace "sched-preemption-3418" to be "running" + Feb 25 10:57:10.538: INFO: Pod "pod1-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 9.277832ms + Feb 25 10:57:10.538: INFO: Pod "pod1-0-sched-preemption-medium-priority" satisfied condition "running" + Feb 25 10:57:10.538: INFO: Waiting up to 5m0s for pod "pod1-1-sched-preemption-medium-priority" in namespace "sched-preemption-3418" to be "running" + Feb 25 10:57:10.547: INFO: Pod "pod1-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 8.79257ms + Feb 25 10:57:10.547: INFO: Pod "pod1-1-sched-preemption-medium-priority" satisfied condition "running" + Feb 25 10:57:10.547: INFO: Waiting up to 5m0s for pod "pod2-0-sched-preemption-medium-priority" in namespace "sched-preemption-3418" to be "running" + Feb 25 10:57:10.554: INFO: Pod "pod2-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 7.76712ms + Feb 25 10:57:10.555: INFO: Pod "pod2-0-sched-preemption-medium-priority" satisfied condition "running" + Feb 25 10:57:10.555: INFO: Waiting up to 5m0s for pod "pod2-1-sched-preemption-medium-priority" in namespace "sched-preemption-3418" to be "running" + Feb 25 10:57:10.562: INFO: Pod "pod2-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 7.036255ms + Feb 25 10:57:10.562: INFO: Pod "pod2-1-sched-preemption-medium-priority" satisfied condition "running" + STEP: Run a critical pod that use same resources as that of a lower priority pod 02/25/23 10:57:10.562 + Feb 25 10:57:10.583: INFO: Waiting up to 2m0s for pod "critical-pod" in namespace "kube-system" to be "running" + Feb 25 10:57:10.596: INFO: Pod "critical-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 12.657902ms + Feb 25 10:57:12.606: INFO: Pod "critical-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022683344s + Feb 25 10:57:14.608: INFO: Pod "critical-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 4.025107863s + Feb 25 10:57:16.606: INFO: Pod "critical-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 6.022833053s + Feb 25 10:57:18.607: INFO: Pod "critical-pod": Phase="Running", Reason="", readiness=true. Elapsed: 8.023858604s + Feb 25 10:57:18.607: INFO: Pod "critical-pod" satisfied condition "running" + [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:187 - Dec 14 16:40:15.059: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "watch-7972" for this suite. 12/14/22 16:40:15.067 + Feb 25 10:57:18.698: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "sched-preemption-3418" for this suite. 02/25/23 10:57:18.717 + [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:80 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSS ------------------------------ -[sig-network] Services - should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance] - test/e2e/network/service.go:2221 -[BeforeEach] [sig-network] Services +[sig-storage] Secrets + should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:124 +[BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:40:15.088 -Dec 14 16:40:15.088: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 16:40:15.089 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:40:15.141 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:40:15.144 -[BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 -[It] should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance] - test/e2e/network/service.go:2221 -STEP: creating service in namespace services-3356 12/14/22 16:40:15.148 -Dec 14 16:40:15.163: INFO: Waiting up to 5m0s for pod "kube-proxy-mode-detector" in namespace "services-3356" to be "running and ready" -Dec 14 16:40:15.175: INFO: Pod "kube-proxy-mode-detector": Phase="Pending", Reason="", readiness=false. Elapsed: 12.581739ms -Dec 14 16:40:15.176: INFO: The phase of Pod kube-proxy-mode-detector is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:40:17.190: INFO: Pod "kube-proxy-mode-detector": Phase="Running", Reason="", readiness=true. Elapsed: 2.026808035s -Dec 14 16:40:17.190: INFO: The phase of Pod kube-proxy-mode-detector is Running (Ready = true) -Dec 14 16:40:17.190: INFO: Pod "kube-proxy-mode-detector" satisfied condition "running and ready" -Dec 14 16:40:17.193: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec kube-proxy-mode-detector -- /bin/sh -x -c curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode' -Dec 14 16:40:17.447: INFO: stderr: "+ curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode\n" -Dec 14 16:40:17.447: INFO: stdout: "iptables" -Dec 14 16:40:17.447: INFO: proxyMode: iptables -Dec 14 16:40:17.465: INFO: Waiting for pod kube-proxy-mode-detector to disappear -Dec 14 16:40:17.469: INFO: Pod kube-proxy-mode-detector no longer exists -STEP: creating service affinity-nodeport-timeout in namespace services-3356 12/14/22 16:40:17.469 -STEP: creating replication controller affinity-nodeport-timeout in namespace services-3356 12/14/22 16:40:17.488 -I1214 16:40:17.498389 14 runners.go:193] Created replication controller with name: affinity-nodeport-timeout, namespace: services-3356, replica count: 3 -I1214 16:40:20.550087 14 runners.go:193] affinity-nodeport-timeout Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -Dec 14 16:40:20.563: INFO: Creating new exec pod -Dec 14 16:40:20.570: INFO: Waiting up to 5m0s for pod "execpod-affinity2hftb" in namespace "services-3356" to be "running" -Dec 14 16:40:20.574: INFO: Pod "execpod-affinity2hftb": Phase="Pending", Reason="", readiness=false. Elapsed: 4.134008ms -Dec 14 16:40:22.582: INFO: Pod "execpod-affinity2hftb": Phase="Running", Reason="", readiness=true. Elapsed: 2.01230833s -Dec 14 16:40:22.582: INFO: Pod "execpod-affinity2hftb" satisfied condition "running" -Dec 14 16:40:23.589: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-nodeport-timeout 80' -Dec 14 16:40:24.190: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-nodeport-timeout 80\nConnection to affinity-nodeport-timeout 80 port [tcp/http] succeeded!\n" -Dec 14 16:40:24.190: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:40:24.190: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.13.178 80' -Dec 14 16:40:24.483: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.13.178 80\nConnection to 10.233.13.178 80 port [tcp/http] succeeded!\n" -Dec 14 16:40:24.483: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:40:24.483: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.249 30257' -Dec 14 16:40:24.782: INFO: stderr: "+ nc -v -t -w 2 192.168.121.249 30257\n+ echo hostName\nConnection to 192.168.121.249 30257 port [tcp/*] succeeded!\n" -Dec 14 16:40:24.782: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:40:24.783: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.197 30257' -Dec 14 16:40:25.099: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.197 30257\nConnection to 192.168.121.197 30257 port [tcp/*] succeeded!\n" -Dec 14 16:40:25.099: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:40:25.099: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.249:30257/ ; done' -Dec 14 16:40:25.475: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n" -Dec 14 16:40:25.475: INFO: stdout: "\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw" -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw -Dec 14 16:40:25.475: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.249:30257/' -Dec 14 16:40:25.705: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n" -Dec 14 16:40:25.705: INFO: stdout: "affinity-nodeport-timeout-fpxxw" -Dec 14 16:40:45.708: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.249:30257/' -Dec 14 16:40:46.208: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n" -Dec 14 16:40:46.208: INFO: stdout: "affinity-nodeport-timeout-fpxxw" -Dec 14 16:41:06.209: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.249:30257/' -Dec 14 16:41:06.548: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n" -Dec 14 16:41:06.548: INFO: stdout: "affinity-nodeport-timeout-fpxxw" -Dec 14 16:41:26.548: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.249:30257/' -Dec 14 16:41:26.826: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n" -Dec 14 16:41:26.826: INFO: stdout: "affinity-nodeport-timeout-phvlv" -Dec 14 16:41:26.826: INFO: Cleaning up the exec pod -STEP: deleting ReplicationController affinity-nodeport-timeout in namespace services-3356, will wait for the garbage collector to delete the pods 12/14/22 16:41:26.862 -Dec 14 16:41:26.932: INFO: Deleting ReplicationController affinity-nodeport-timeout took: 13.399682ms -Dec 14 16:41:27.033: INFO: Terminating ReplicationController affinity-nodeport-timeout pods took: 101.08608ms -[AfterEach] [sig-network] Services +STEP: Creating a kubernetes client 02/25/23 10:57:18.876 +Feb 25 10:57:18.876: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename secrets 02/25/23 10:57:18.879 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:57:18.977 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:57:19 +[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:124 +STEP: Creating secret with name secret-test-c8694d66-bc40-4f30-aea1-f07c920f98b0 02/25/23 10:57:19.013 +STEP: Creating a pod to test consume secrets 02/25/23 10:57:19.027 +Feb 25 10:57:19.155: INFO: Waiting up to 5m0s for pod "pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397" in namespace "secrets-9055" to be "Succeeded or Failed" +Feb 25 10:57:19.166: INFO: Pod "pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397": Phase="Pending", Reason="", readiness=false. Elapsed: 11.287517ms +Feb 25 10:57:21.190: INFO: Pod "pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397": Phase="Pending", Reason="", readiness=false. Elapsed: 2.035174062s +Feb 25 10:57:23.187: INFO: Pod "pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397": Phase="Pending", Reason="", readiness=false. Elapsed: 4.032179426s +Feb 25 10:57:25.202: INFO: Pod "pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.047817127s +STEP: Saw pod success 02/25/23 10:57:25.203 +Feb 25 10:57:25.203: INFO: Pod "pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397" satisfied condition "Succeeded or Failed" +Feb 25 10:57:25.238: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397 container secret-volume-test: +STEP: delete the pod 02/25/23 10:57:25.335 +Feb 25 10:57:25.455: INFO: Waiting for pod pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397 to disappear +Feb 25 10:57:25.479: INFO: Pod pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397 no longer exists +[AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 -Dec 14 16:41:29.110: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-3356" for this suite. 12/14/22 16:41:29.12 -[AfterEach] [sig-network] Services - test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance]","completed":216,"skipped":4035,"failed":0} +Feb 25 10:57:25.480: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-9055" for this suite. 02/25/23 10:57:25.556 +{"msg":"PASSED [sig-storage] Secrets should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","completed":213,"skipped":3927,"failed":0} ------------------------------ -• [SLOW TEST] [74.040 seconds] -[sig-network] Services -test/e2e/network/common/framework.go:23 - should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance] - test/e2e/network/service.go:2221 +• [SLOW TEST] [6.706 seconds] +[sig-storage] Secrets +test/e2e/common/storage/framework.go:23 + should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:124 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Services + [BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:40:15.088 - Dec 14 16:40:15.088: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 16:40:15.089 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:40:15.141 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:40:15.144 - [BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 - [It] should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance] - test/e2e/network/service.go:2221 - STEP: creating service in namespace services-3356 12/14/22 16:40:15.148 - Dec 14 16:40:15.163: INFO: Waiting up to 5m0s for pod "kube-proxy-mode-detector" in namespace "services-3356" to be "running and ready" - Dec 14 16:40:15.175: INFO: Pod "kube-proxy-mode-detector": Phase="Pending", Reason="", readiness=false. Elapsed: 12.581739ms - Dec 14 16:40:15.176: INFO: The phase of Pod kube-proxy-mode-detector is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:40:17.190: INFO: Pod "kube-proxy-mode-detector": Phase="Running", Reason="", readiness=true. Elapsed: 2.026808035s - Dec 14 16:40:17.190: INFO: The phase of Pod kube-proxy-mode-detector is Running (Ready = true) - Dec 14 16:40:17.190: INFO: Pod "kube-proxy-mode-detector" satisfied condition "running and ready" - Dec 14 16:40:17.193: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec kube-proxy-mode-detector -- /bin/sh -x -c curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode' - Dec 14 16:40:17.447: INFO: stderr: "+ curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode\n" - Dec 14 16:40:17.447: INFO: stdout: "iptables" - Dec 14 16:40:17.447: INFO: proxyMode: iptables - Dec 14 16:40:17.465: INFO: Waiting for pod kube-proxy-mode-detector to disappear - Dec 14 16:40:17.469: INFO: Pod kube-proxy-mode-detector no longer exists - STEP: creating service affinity-nodeport-timeout in namespace services-3356 12/14/22 16:40:17.469 - STEP: creating replication controller affinity-nodeport-timeout in namespace services-3356 12/14/22 16:40:17.488 - I1214 16:40:17.498389 14 runners.go:193] Created replication controller with name: affinity-nodeport-timeout, namespace: services-3356, replica count: 3 - I1214 16:40:20.550087 14 runners.go:193] affinity-nodeport-timeout Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - Dec 14 16:40:20.563: INFO: Creating new exec pod - Dec 14 16:40:20.570: INFO: Waiting up to 5m0s for pod "execpod-affinity2hftb" in namespace "services-3356" to be "running" - Dec 14 16:40:20.574: INFO: Pod "execpod-affinity2hftb": Phase="Pending", Reason="", readiness=false. Elapsed: 4.134008ms - Dec 14 16:40:22.582: INFO: Pod "execpod-affinity2hftb": Phase="Running", Reason="", readiness=true. Elapsed: 2.01230833s - Dec 14 16:40:22.582: INFO: Pod "execpod-affinity2hftb" satisfied condition "running" - Dec 14 16:40:23.589: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-nodeport-timeout 80' - Dec 14 16:40:24.190: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-nodeport-timeout 80\nConnection to affinity-nodeport-timeout 80 port [tcp/http] succeeded!\n" - Dec 14 16:40:24.190: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:40:24.190: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.13.178 80' - Dec 14 16:40:24.483: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.13.178 80\nConnection to 10.233.13.178 80 port [tcp/http] succeeded!\n" - Dec 14 16:40:24.483: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:40:24.483: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.249 30257' - Dec 14 16:40:24.782: INFO: stderr: "+ nc -v -t -w 2 192.168.121.249 30257\n+ echo hostName\nConnection to 192.168.121.249 30257 port [tcp/*] succeeded!\n" - Dec 14 16:40:24.782: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:40:24.783: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.197 30257' - Dec 14 16:40:25.099: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.197 30257\nConnection to 192.168.121.197 30257 port [tcp/*] succeeded!\n" - Dec 14 16:40:25.099: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:40:25.099: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.249:30257/ ; done' - Dec 14 16:40:25.475: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n" - Dec 14 16:40:25.475: INFO: stdout: "\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw\naffinity-nodeport-timeout-fpxxw" - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Received response from host: affinity-nodeport-timeout-fpxxw - Dec 14 16:40:25.475: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.249:30257/' - Dec 14 16:40:25.705: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n" - Dec 14 16:40:25.705: INFO: stdout: "affinity-nodeport-timeout-fpxxw" - Dec 14 16:40:45.708: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.249:30257/' - Dec 14 16:40:46.208: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n" - Dec 14 16:40:46.208: INFO: stdout: "affinity-nodeport-timeout-fpxxw" - Dec 14 16:41:06.209: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.249:30257/' - Dec 14 16:41:06.548: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n" - Dec 14 16:41:06.548: INFO: stdout: "affinity-nodeport-timeout-fpxxw" - Dec 14 16:41:26.548: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3356 exec execpod-affinity2hftb -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.249:30257/' - Dec 14 16:41:26.826: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.249:30257/\n" - Dec 14 16:41:26.826: INFO: stdout: "affinity-nodeport-timeout-phvlv" - Dec 14 16:41:26.826: INFO: Cleaning up the exec pod - STEP: deleting ReplicationController affinity-nodeport-timeout in namespace services-3356, will wait for the garbage collector to delete the pods 12/14/22 16:41:26.862 - Dec 14 16:41:26.932: INFO: Deleting ReplicationController affinity-nodeport-timeout took: 13.399682ms - Dec 14 16:41:27.033: INFO: Terminating ReplicationController affinity-nodeport-timeout pods took: 101.08608ms - [AfterEach] [sig-network] Services + STEP: Creating a kubernetes client 02/25/23 10:57:18.876 + Feb 25 10:57:18.876: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename secrets 02/25/23 10:57:18.879 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:57:18.977 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:57:19 + [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:124 + STEP: Creating secret with name secret-test-c8694d66-bc40-4f30-aea1-f07c920f98b0 02/25/23 10:57:19.013 + STEP: Creating a pod to test consume secrets 02/25/23 10:57:19.027 + Feb 25 10:57:19.155: INFO: Waiting up to 5m0s for pod "pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397" in namespace "secrets-9055" to be "Succeeded or Failed" + Feb 25 10:57:19.166: INFO: Pod "pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397": Phase="Pending", Reason="", readiness=false. Elapsed: 11.287517ms + Feb 25 10:57:21.190: INFO: Pod "pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397": Phase="Pending", Reason="", readiness=false. Elapsed: 2.035174062s + Feb 25 10:57:23.187: INFO: Pod "pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397": Phase="Pending", Reason="", readiness=false. Elapsed: 4.032179426s + Feb 25 10:57:25.202: INFO: Pod "pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.047817127s + STEP: Saw pod success 02/25/23 10:57:25.203 + Feb 25 10:57:25.203: INFO: Pod "pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397" satisfied condition "Succeeded or Failed" + Feb 25 10:57:25.238: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397 container secret-volume-test: + STEP: delete the pod 02/25/23 10:57:25.335 + Feb 25 10:57:25.455: INFO: Waiting for pod pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397 to disappear + Feb 25 10:57:25.479: INFO: Pod pod-secrets-f055bd11-86d1-4c9b-86b1-0fb77f736397 no longer exists + [AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 - Dec 14 16:41:29.110: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-3356" for this suite. 12/14/22 16:41:29.12 - [AfterEach] [sig-network] Services - test/e2e/network/service.go:762 + Feb 25 10:57:25.480: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "secrets-9055" for this suite. 02/25/23 10:57:25.556 << End Captured GinkgoWriter Output ------------------------------ -S +SSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:106 -[BeforeEach] [sig-storage] EmptyDir volumes +[sig-node] Container Runtime blackbox test on terminated container + should report termination message from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:215 +[BeforeEach] [sig-node] Container Runtime test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:41:29.128 -Dec 14 16:41:29.129: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir 12/14/22 16:41:29.131 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:29.176 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:29.179 -[It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:106 -STEP: Creating a pod to test emptydir 0666 on tmpfs 12/14/22 16:41:29.185 -Dec 14 16:41:29.198: INFO: Waiting up to 5m0s for pod "pod-79674faa-6bb0-4111-9a07-a09970ed7d9d" in namespace "emptydir-9898" to be "Succeeded or Failed" -Dec 14 16:41:29.209: INFO: Pod "pod-79674faa-6bb0-4111-9a07-a09970ed7d9d": Phase="Pending", Reason="", readiness=false. Elapsed: 10.30295ms -Dec 14 16:41:31.215: INFO: Pod "pod-79674faa-6bb0-4111-9a07-a09970ed7d9d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016759727s -Dec 14 16:41:33.214: INFO: Pod "pod-79674faa-6bb0-4111-9a07-a09970ed7d9d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015715034s -STEP: Saw pod success 12/14/22 16:41:33.214 -Dec 14 16:41:33.215: INFO: Pod "pod-79674faa-6bb0-4111-9a07-a09970ed7d9d" satisfied condition "Succeeded or Failed" -Dec 14 16:41:33.218: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-79674faa-6bb0-4111-9a07-a09970ed7d9d container test-container: -STEP: delete the pod 12/14/22 16:41:33.227 -Dec 14 16:41:33.243: INFO: Waiting for pod pod-79674faa-6bb0-4111-9a07-a09970ed7d9d to disappear -Dec 14 16:41:33.247: INFO: Pod pod-79674faa-6bb0-4111-9a07-a09970ed7d9d no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +STEP: Creating a kubernetes client 02/25/23 10:57:25.584 +Feb 25 10:57:25.585: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-runtime 02/25/23 10:57:25.605 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:57:25.688 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:57:25.719 +[It] should report termination message from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:215 +STEP: create the container 02/25/23 10:57:25.734 +STEP: wait for the container to reach Failed 02/25/23 10:57:25.776 +STEP: get the container status 02/25/23 10:57:33.925 +STEP: the container should be terminated 02/25/23 10:57:33.932 +STEP: the termination message should be set 02/25/23 10:57:33.932 +Feb 25 10:57:33.933: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- +STEP: delete the container 02/25/23 10:57:33.933 +[AfterEach] [sig-node] Container Runtime test/e2e/framework/framework.go:187 -Dec 14 16:41:33.247: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-9898" for this suite. 12/14/22 16:41:33.253 -{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","completed":217,"skipped":4036,"failed":0} +Feb 25 10:57:33.964: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-runtime-8459" for this suite. 02/25/23 10:57:33.973 +{"msg":"PASSED [sig-node] Container Runtime blackbox test on terminated container should report termination message from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","completed":214,"skipped":3934,"failed":0} ------------------------------ -• [4.134 seconds] -[sig-storage] EmptyDir volumes -test/e2e/common/storage/framework.go:23 - should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:106 +• [SLOW TEST] [8.417 seconds] +[sig-node] Container Runtime +test/e2e/common/node/framework.go:23 + blackbox test + test/e2e/common/node/runtime.go:43 + on terminated container + test/e2e/common/node/runtime.go:136 + should report termination message from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:215 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir volumes + [BeforeEach] [sig-node] Container Runtime test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:41:29.128 - Dec 14 16:41:29.129: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir 12/14/22 16:41:29.131 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:29.176 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:29.179 - [It] should support (root,0666,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:106 - STEP: Creating a pod to test emptydir 0666 on tmpfs 12/14/22 16:41:29.185 - Dec 14 16:41:29.198: INFO: Waiting up to 5m0s for pod "pod-79674faa-6bb0-4111-9a07-a09970ed7d9d" in namespace "emptydir-9898" to be "Succeeded or Failed" - Dec 14 16:41:29.209: INFO: Pod "pod-79674faa-6bb0-4111-9a07-a09970ed7d9d": Phase="Pending", Reason="", readiness=false. Elapsed: 10.30295ms - Dec 14 16:41:31.215: INFO: Pod "pod-79674faa-6bb0-4111-9a07-a09970ed7d9d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016759727s - Dec 14 16:41:33.214: INFO: Pod "pod-79674faa-6bb0-4111-9a07-a09970ed7d9d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.015715034s - STEP: Saw pod success 12/14/22 16:41:33.214 - Dec 14 16:41:33.215: INFO: Pod "pod-79674faa-6bb0-4111-9a07-a09970ed7d9d" satisfied condition "Succeeded or Failed" - Dec 14 16:41:33.218: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-79674faa-6bb0-4111-9a07-a09970ed7d9d container test-container: - STEP: delete the pod 12/14/22 16:41:33.227 - Dec 14 16:41:33.243: INFO: Waiting for pod pod-79674faa-6bb0-4111-9a07-a09970ed7d9d to disappear - Dec 14 16:41:33.247: INFO: Pod pod-79674faa-6bb0-4111-9a07-a09970ed7d9d no longer exists - [AfterEach] [sig-storage] EmptyDir volumes + STEP: Creating a kubernetes client 02/25/23 10:57:25.584 + Feb 25 10:57:25.585: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-runtime 02/25/23 10:57:25.605 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:57:25.688 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:57:25.719 + [It] should report termination message from log output if TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:215 + STEP: create the container 02/25/23 10:57:25.734 + STEP: wait for the container to reach Failed 02/25/23 10:57:25.776 + STEP: get the container status 02/25/23 10:57:33.925 + STEP: the container should be terminated 02/25/23 10:57:33.932 + STEP: the termination message should be set 02/25/23 10:57:33.932 + Feb 25 10:57:33.933: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- + STEP: delete the container 02/25/23 10:57:33.933 + [AfterEach] [sig-node] Container Runtime test/e2e/framework/framework.go:187 - Dec 14 16:41:33.247: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-9898" for this suite. 12/14/22 16:41:33.253 + Feb 25 10:57:33.964: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-runtime-8459" for this suite. 02/25/23 10:57:33.973 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSS ------------------------------ -[sig-cli] Kubectl client Kubectl api-versions - should check if v1 is in available api versions [Conformance] - test/e2e/kubectl/kubectl.go:822 +[sig-cli] Kubectl client Kubectl server-side dry-run + should check if kubectl can dry-run update Pods [Conformance] + test/e2e/kubectl/kubectl.go:960 [BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:41:33.27 -Dec 14 16:41:33.270: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 16:41:33.271 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:33.302 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:33.306 +STEP: Creating a kubernetes client 02/25/23 10:57:34.008 +Feb 25 10:57:34.008: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 10:57:34.01 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:57:34.044 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:57:34.05 [BeforeEach] [sig-cli] Kubectl client test/e2e/kubectl/kubectl.go:272 -[It] should check if v1 is in available api versions [Conformance] - test/e2e/kubectl/kubectl.go:822 -STEP: validating api versions 12/14/22 16:41:33.309 -Dec 14 16:41:33.310: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3307 api-versions' -Dec 14 16:41:33.415: INFO: stderr: "" -Dec 14 16:41:33.415: INFO: stdout: "admissionregistration.k8s.io/v1\napiextensions.k8s.io/v1\napiregistration.k8s.io/v1\napps/v1\nauthentication.k8s.io/v1\nauthorization.k8s.io/v1\nautoscaling/v1\nautoscaling/v2\nautoscaling/v2beta2\nbatch/v1\ncertificates.k8s.io/v1\ncoordination.k8s.io/v1\ndiscovery.k8s.io/v1\nevents.k8s.io/v1\nflowcontrol.apiserver.k8s.io/v1beta1\nflowcontrol.apiserver.k8s.io/v1beta2\nnetworking.k8s.io/v1\nnode.k8s.io/v1\npolicy/v1\nrbac.authorization.k8s.io/v1\nscheduling.k8s.io/v1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n" +[It] should check if kubectl can dry-run update Pods [Conformance] + test/e2e/kubectl/kubectl.go:960 +STEP: running the image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 02/25/23 10:57:34.059 +Feb 25 10:57:34.060: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1336 run e2e-test-httpd-pod --image=registry.k8s.io/e2e-test-images/httpd:2.4.38-2 --pod-running-timeout=2m0s --labels=run=e2e-test-httpd-pod' +Feb 25 10:57:34.402: INFO: stderr: "" +Feb 25 10:57:34.404: INFO: stdout: "pod/e2e-test-httpd-pod created\n" +STEP: replace the image in the pod with server-side dry-run 02/25/23 10:57:34.404 +Feb 25 10:57:34.407: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1336 patch pod e2e-test-httpd-pod -p {"spec":{"containers":[{"name": "e2e-test-httpd-pod","image": "registry.k8s.io/e2e-test-images/busybox:1.29-2"}]}} --dry-run=server' +Feb 25 10:57:36.558: INFO: stderr: "" +Feb 25 10:57:36.558: INFO: stdout: "pod/e2e-test-httpd-pod patched\n" +STEP: verifying the pod e2e-test-httpd-pod has the right image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 02/25/23 10:57:36.558 +Feb 25 10:57:36.579: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1336 delete pods e2e-test-httpd-pod' +Feb 25 10:57:39.655: INFO: stderr: "" +Feb 25 10:57:39.655: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 -Dec 14 16:41:33.415: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-3307" for this suite. 12/14/22 16:41:33.42 -{"msg":"PASSED [sig-cli] Kubectl client Kubectl api-versions should check if v1 is in available api versions [Conformance]","completed":218,"skipped":4069,"failed":0} +Feb 25 10:57:39.655: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-1336" for this suite. 02/25/23 10:57:39.67 +{"msg":"PASSED [sig-cli] Kubectl client Kubectl server-side dry-run should check if kubectl can dry-run update Pods [Conformance]","completed":215,"skipped":3945,"failed":0} ------------------------------ -• [0.164 seconds] +• [SLOW TEST] [5.690 seconds] [sig-cli] Kubectl client test/e2e/kubectl/framework.go:23 - Kubectl api-versions - test/e2e/kubectl/kubectl.go:816 - should check if v1 is in available api versions [Conformance] - test/e2e/kubectl/kubectl.go:822 + Kubectl server-side dry-run + test/e2e/kubectl/kubectl.go:954 + should check if kubectl can dry-run update Pods [Conformance] + test/e2e/kubectl/kubectl.go:960 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:41:33.27 - Dec 14 16:41:33.270: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 16:41:33.271 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:33.302 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:33.306 + STEP: Creating a kubernetes client 02/25/23 10:57:34.008 + Feb 25 10:57:34.008: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 10:57:34.01 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:57:34.044 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:57:34.05 [BeforeEach] [sig-cli] Kubectl client test/e2e/kubectl/kubectl.go:272 - [It] should check if v1 is in available api versions [Conformance] - test/e2e/kubectl/kubectl.go:822 - STEP: validating api versions 12/14/22 16:41:33.309 - Dec 14 16:41:33.310: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3307 api-versions' - Dec 14 16:41:33.415: INFO: stderr: "" - Dec 14 16:41:33.415: INFO: stdout: "admissionregistration.k8s.io/v1\napiextensions.k8s.io/v1\napiregistration.k8s.io/v1\napps/v1\nauthentication.k8s.io/v1\nauthorization.k8s.io/v1\nautoscaling/v1\nautoscaling/v2\nautoscaling/v2beta2\nbatch/v1\ncertificates.k8s.io/v1\ncoordination.k8s.io/v1\ndiscovery.k8s.io/v1\nevents.k8s.io/v1\nflowcontrol.apiserver.k8s.io/v1beta1\nflowcontrol.apiserver.k8s.io/v1beta2\nnetworking.k8s.io/v1\nnode.k8s.io/v1\npolicy/v1\nrbac.authorization.k8s.io/v1\nscheduling.k8s.io/v1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n" + [It] should check if kubectl can dry-run update Pods [Conformance] + test/e2e/kubectl/kubectl.go:960 + STEP: running the image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 02/25/23 10:57:34.059 + Feb 25 10:57:34.060: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1336 run e2e-test-httpd-pod --image=registry.k8s.io/e2e-test-images/httpd:2.4.38-2 --pod-running-timeout=2m0s --labels=run=e2e-test-httpd-pod' + Feb 25 10:57:34.402: INFO: stderr: "" + Feb 25 10:57:34.404: INFO: stdout: "pod/e2e-test-httpd-pod created\n" + STEP: replace the image in the pod with server-side dry-run 02/25/23 10:57:34.404 + Feb 25 10:57:34.407: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1336 patch pod e2e-test-httpd-pod -p {"spec":{"containers":[{"name": "e2e-test-httpd-pod","image": "registry.k8s.io/e2e-test-images/busybox:1.29-2"}]}} --dry-run=server' + Feb 25 10:57:36.558: INFO: stderr: "" + Feb 25 10:57:36.558: INFO: stdout: "pod/e2e-test-httpd-pod patched\n" + STEP: verifying the pod e2e-test-httpd-pod has the right image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 02/25/23 10:57:36.558 + Feb 25 10:57:36.579: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1336 delete pods e2e-test-httpd-pod' + Feb 25 10:57:39.655: INFO: stderr: "" + Feb 25 10:57:39.655: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 16:41:33.415: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-3307" for this suite. 12/14/22 16:41:33.42 + Feb 25 10:57:39.655: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-1336" for this suite. 02/25/23 10:57:39.67 << End Captured GinkgoWriter Output ------------------------------ -SSS +SS ------------------------------ -[sig-cli] Kubectl client Kubectl patch - should add annotations for pods in rc [Conformance] - test/e2e/kubectl/kubectl.go:1650 -[BeforeEach] [sig-cli] Kubectl client +[sig-scheduling] SchedulerPreemption [Serial] PriorityClass endpoints + verify PriorityClass endpoints can be operated with different HTTP methods [Conformance] + test/e2e/scheduling/preemption.go:733 +[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:41:33.436 -Dec 14 16:41:33.436: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 16:41:33.438 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:33.459 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:33.462 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 -[It] should add annotations for pods in rc [Conformance] - test/e2e/kubectl/kubectl.go:1650 -STEP: creating Agnhost RC 12/14/22 16:41:33.465 -Dec 14 16:41:33.466: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2919 create -f -' -Dec 14 16:41:34.753: INFO: stderr: "" -Dec 14 16:41:34.753: INFO: stdout: "replicationcontroller/agnhost-primary created\n" -STEP: Waiting for Agnhost primary to start. 12/14/22 16:41:34.753 -Dec 14 16:41:35.770: INFO: Selector matched 1 pods for map[app:agnhost] -Dec 14 16:41:35.770: INFO: Found 1 / 1 -Dec 14 16:41:35.770: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 -STEP: patching all pods 12/14/22 16:41:35.77 -Dec 14 16:41:35.775: INFO: Selector matched 1 pods for map[app:agnhost] -Dec 14 16:41:35.775: INFO: ForEach: Found 1 pods from the filter. Now looping through them. -Dec 14 16:41:35.775: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2919 patch pod agnhost-primary-l7s4v -p {"metadata":{"annotations":{"x":"y"}}}' -Dec 14 16:41:35.916: INFO: stderr: "" -Dec 14 16:41:35.916: INFO: stdout: "pod/agnhost-primary-l7s4v patched\n" -STEP: checking annotations 12/14/22 16:41:35.916 -Dec 14 16:41:35.920: INFO: Selector matched 1 pods for map[app:agnhost] -Dec 14 16:41:35.921: INFO: ForEach: Found 1 pods from the filter. Now looping through them. -[AfterEach] [sig-cli] Kubectl client +STEP: Creating a kubernetes client 02/25/23 10:57:39.703 +Feb 25 10:57:39.703: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename sched-preemption 02/25/23 10:57:39.706 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:57:39.808 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:57:39.814 +[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:92 +Feb 25 10:57:39.876: INFO: Waiting up to 1m0s for all nodes to be ready +Feb 25 10:58:39.940: INFO: Waiting for terminating namespaces to be deleted... +[BeforeEach] PriorityClass endpoints + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 10:58:39.948 +Feb 25 10:58:39.949: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename sched-preemption-path 02/25/23 10:58:39.953 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:58:39.987 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:58:39.996 +[BeforeEach] PriorityClass endpoints + test/e2e/scheduling/preemption.go:690 +[It] verify PriorityClass endpoints can be operated with different HTTP methods [Conformance] + test/e2e/scheduling/preemption.go:733 +Feb 25 10:58:40.038: INFO: PriorityClass.scheduling.k8s.io "p1" is invalid: value: Forbidden: may not be changed in an update. +Feb 25 10:58:40.047: INFO: PriorityClass.scheduling.k8s.io "p2" is invalid: value: Forbidden: may not be changed in an update. +[AfterEach] PriorityClass endpoints + test/e2e/framework/framework.go:187 +Feb 25 10:58:40.094: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "sched-preemption-path-2623" for this suite. 02/25/23 10:58:40.107 +[AfterEach] PriorityClass endpoints + test/e2e/scheduling/preemption.go:706 +[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:187 -Dec 14 16:41:35.921: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-2919" for this suite. 12/14/22 16:41:35.926 -{"msg":"PASSED [sig-cli] Kubectl client Kubectl patch should add annotations for pods in rc [Conformance]","completed":219,"skipped":4072,"failed":0} +Feb 25 10:58:40.155: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "sched-preemption-2626" for this suite. 02/25/23 10:58:40.166 +[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:80 +{"msg":"PASSED [sig-scheduling] SchedulerPreemption [Serial] PriorityClass endpoints verify PriorityClass endpoints can be operated with different HTTP methods [Conformance]","completed":216,"skipped":3947,"failed":0} ------------------------------ -• [2.497 seconds] -[sig-cli] Kubectl client -test/e2e/kubectl/framework.go:23 - Kubectl patch - test/e2e/kubectl/kubectl.go:1644 - should add annotations for pods in rc [Conformance] - test/e2e/kubectl/kubectl.go:1650 +• [SLOW TEST] [60.602 seconds] +[sig-scheduling] SchedulerPreemption [Serial] +test/e2e/scheduling/framework.go:40 + PriorityClass endpoints + test/e2e/scheduling/preemption.go:683 + verify PriorityClass endpoints can be operated with different HTTP methods [Conformance] + test/e2e/scheduling/preemption.go:733 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-cli] Kubectl client + [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:41:33.436 - Dec 14 16:41:33.436: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 16:41:33.438 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:33.459 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:33.462 - [BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 - [It] should add annotations for pods in rc [Conformance] - test/e2e/kubectl/kubectl.go:1650 - STEP: creating Agnhost RC 12/14/22 16:41:33.465 - Dec 14 16:41:33.466: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2919 create -f -' - Dec 14 16:41:34.753: INFO: stderr: "" - Dec 14 16:41:34.753: INFO: stdout: "replicationcontroller/agnhost-primary created\n" - STEP: Waiting for Agnhost primary to start. 12/14/22 16:41:34.753 - Dec 14 16:41:35.770: INFO: Selector matched 1 pods for map[app:agnhost] - Dec 14 16:41:35.770: INFO: Found 1 / 1 - Dec 14 16:41:35.770: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 - STEP: patching all pods 12/14/22 16:41:35.77 - Dec 14 16:41:35.775: INFO: Selector matched 1 pods for map[app:agnhost] - Dec 14 16:41:35.775: INFO: ForEach: Found 1 pods from the filter. Now looping through them. - Dec 14 16:41:35.775: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2919 patch pod agnhost-primary-l7s4v -p {"metadata":{"annotations":{"x":"y"}}}' - Dec 14 16:41:35.916: INFO: stderr: "" - Dec 14 16:41:35.916: INFO: stdout: "pod/agnhost-primary-l7s4v patched\n" - STEP: checking annotations 12/14/22 16:41:35.916 - Dec 14 16:41:35.920: INFO: Selector matched 1 pods for map[app:agnhost] - Dec 14 16:41:35.921: INFO: ForEach: Found 1 pods from the filter. Now looping through them. - [AfterEach] [sig-cli] Kubectl client + STEP: Creating a kubernetes client 02/25/23 10:57:39.703 + Feb 25 10:57:39.703: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename sched-preemption 02/25/23 10:57:39.706 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:57:39.808 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:57:39.814 + [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:92 + Feb 25 10:57:39.876: INFO: Waiting up to 1m0s for all nodes to be ready + Feb 25 10:58:39.940: INFO: Waiting for terminating namespaces to be deleted... + [BeforeEach] PriorityClass endpoints + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 10:58:39.948 + Feb 25 10:58:39.949: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename sched-preemption-path 02/25/23 10:58:39.953 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:58:39.987 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:58:39.996 + [BeforeEach] PriorityClass endpoints + test/e2e/scheduling/preemption.go:690 + [It] verify PriorityClass endpoints can be operated with different HTTP methods [Conformance] + test/e2e/scheduling/preemption.go:733 + Feb 25 10:58:40.038: INFO: PriorityClass.scheduling.k8s.io "p1" is invalid: value: Forbidden: may not be changed in an update. + Feb 25 10:58:40.047: INFO: PriorityClass.scheduling.k8s.io "p2" is invalid: value: Forbidden: may not be changed in an update. + [AfterEach] PriorityClass endpoints + test/e2e/framework/framework.go:187 + Feb 25 10:58:40.094: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "sched-preemption-path-2623" for this suite. 02/25/23 10:58:40.107 + [AfterEach] PriorityClass endpoints + test/e2e/scheduling/preemption.go:706 + [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:187 - Dec 14 16:41:35.921: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-2919" for this suite. 12/14/22 16:41:35.926 + Feb 25 10:58:40.155: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "sched-preemption-2626" for this suite. 02/25/23 10:58:40.166 + [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:80 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-cli] Kubectl client Kubectl replace - should update a single-container pod's image [Conformance] - test/e2e/kubectl/kubectl.go:1745 +[sig-cli] Kubectl client Update Demo + should scale a replication controller [Conformance] + test/e2e/kubectl/kubectl.go:350 [BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:41:35.935 -Dec 14 16:41:35.935: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 16:41:35.939 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:35.965 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:35.969 +STEP: Creating a kubernetes client 02/25/23 10:58:40.321 +Feb 25 10:58:40.321: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 10:58:40.324 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:58:40.352 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:58:40.36 [BeforeEach] [sig-cli] Kubectl client test/e2e/kubectl/kubectl.go:272 -[BeforeEach] Kubectl replace - test/e2e/kubectl/kubectl.go:1732 -[It] should update a single-container pod's image [Conformance] - test/e2e/kubectl/kubectl.go:1745 -STEP: running the image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 12/14/22 16:41:35.972 -Dec 14 16:41:35.972: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2615 run e2e-test-httpd-pod --image=registry.k8s.io/e2e-test-images/httpd:2.4.38-2 --pod-running-timeout=2m0s --labels=run=e2e-test-httpd-pod' -Dec 14 16:41:36.155: INFO: stderr: "" -Dec 14 16:41:36.156: INFO: stdout: "pod/e2e-test-httpd-pod created\n" -STEP: verifying the pod e2e-test-httpd-pod is running 12/14/22 16:41:36.156 -STEP: verifying the pod e2e-test-httpd-pod was created 12/14/22 16:41:41.208 -Dec 14 16:41:41.209: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2615 get pod e2e-test-httpd-pod -o json' -Dec 14 16:41:41.337: INFO: stderr: "" -Dec 14 16:41:41.337: INFO: stdout: "{\n \"apiVersion\": \"v1\",\n \"kind\": \"Pod\",\n \"metadata\": {\n \"creationTimestamp\": \"2022-12-14T16:41:36Z\",\n \"labels\": {\n \"run\": \"e2e-test-httpd-pod\"\n },\n \"name\": \"e2e-test-httpd-pod\",\n \"namespace\": \"kubectl-2615\",\n \"resourceVersion\": \"22676\",\n \"uid\": \"b5c068db-1a16-4c2f-ab3a-2dc68c28e44e\"\n },\n \"spec\": {\n \"containers\": [\n {\n \"image\": \"registry.k8s.io/e2e-test-images/httpd:2.4.38-2\",\n \"imagePullPolicy\": \"IfNotPresent\",\n \"name\": \"e2e-test-httpd-pod\",\n \"resources\": {},\n \"terminationMessagePath\": \"/dev/termination-log\",\n \"terminationMessagePolicy\": \"File\",\n \"volumeMounts\": [\n {\n \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n \"name\": \"kube-api-access-nq7k2\",\n \"readOnly\": true\n }\n ]\n }\n ],\n \"dnsPolicy\": \"ClusterFirst\",\n \"enableServiceLinks\": true,\n \"nodeName\": \"kiedai4iakae-3\",\n \"preemptionPolicy\": \"PreemptLowerPriority\",\n \"priority\": 0,\n \"restartPolicy\": \"Always\",\n \"schedulerName\": \"default-scheduler\",\n \"securityContext\": {},\n \"serviceAccount\": \"default\",\n \"serviceAccountName\": \"default\",\n \"terminationGracePeriodSeconds\": 30,\n \"tolerations\": [\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/not-ready\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n },\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/unreachable\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n }\n ],\n \"volumes\": [\n {\n \"name\": \"kube-api-access-nq7k2\",\n \"projected\": {\n \"defaultMode\": 420,\n \"sources\": [\n {\n \"serviceAccountToken\": {\n \"expirationSeconds\": 3607,\n \"path\": \"token\"\n }\n },\n {\n \"configMap\": {\n \"items\": [\n {\n \"key\": \"ca.crt\",\n \"path\": \"ca.crt\"\n }\n ],\n \"name\": \"kube-root-ca.crt\"\n }\n },\n {\n \"downwardAPI\": {\n \"items\": [\n {\n \"fieldRef\": {\n \"apiVersion\": \"v1\",\n \"fieldPath\": \"metadata.namespace\"\n },\n \"path\": \"namespace\"\n }\n ]\n }\n }\n ]\n }\n }\n ]\n },\n \"status\": {\n \"conditions\": [\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2022-12-14T16:41:36Z\",\n \"status\": \"True\",\n \"type\": \"Initialized\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2022-12-14T16:41:37Z\",\n \"status\": \"True\",\n \"type\": \"Ready\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2022-12-14T16:41:37Z\",\n \"status\": \"True\",\n \"type\": \"ContainersReady\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2022-12-14T16:41:36Z\",\n \"status\": \"True\",\n \"type\": \"PodScheduled\"\n }\n ],\n \"containerStatuses\": [\n {\n \"containerID\": \"cri-o://1d429c9067a346e37eb8a6cb447ed6132b96763a0ea07f1c8c42d040e5d3d23b\",\n \"image\": \"registry.k8s.io/e2e-test-images/httpd:2.4.38-2\",\n \"imageID\": \"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3\",\n \"lastState\": {},\n \"name\": \"e2e-test-httpd-pod\",\n \"ready\": true,\n \"restartCount\": 0,\n \"started\": true,\n \"state\": {\n \"running\": {\n \"startedAt\": \"2022-12-14T16:41:37Z\"\n }\n }\n }\n ],\n \"hostIP\": \"192.168.121.244\",\n \"phase\": \"Running\",\n \"podIP\": \"10.233.66.60\",\n \"podIPs\": [\n {\n \"ip\": \"10.233.66.60\"\n }\n ],\n \"qosClass\": \"BestEffort\",\n \"startTime\": \"2022-12-14T16:41:36Z\"\n }\n}\n" -STEP: replace the image in the pod 12/14/22 16:41:41.338 -Dec 14 16:41:41.338: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2615 replace -f -' -Dec 14 16:41:41.664: INFO: stderr: "" -Dec 14 16:41:41.664: INFO: stdout: "pod/e2e-test-httpd-pod replaced\n" -STEP: verifying the pod e2e-test-httpd-pod has the right image registry.k8s.io/e2e-test-images/busybox:1.29-2 12/14/22 16:41:41.665 -[AfterEach] Kubectl replace - test/e2e/kubectl/kubectl.go:1736 -Dec 14 16:41:41.694: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2615 delete pods e2e-test-httpd-pod' -Dec 14 16:41:43.753: INFO: stderr: "" -Dec 14 16:41:43.753: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" +[BeforeEach] Update Demo + test/e2e/kubectl/kubectl.go:324 +[It] should scale a replication controller [Conformance] + test/e2e/kubectl/kubectl.go:350 +STEP: creating a replication controller 02/25/23 10:58:40.368 +Feb 25 10:58:40.369: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 create -f -' +Feb 25 10:58:41.333: INFO: stderr: "" +Feb 25 10:58:41.333: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" +STEP: waiting for all containers in name=update-demo pods to come up. 02/25/23 10:58:41.333 +Feb 25 10:58:41.335: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' +Feb 25 10:58:41.623: INFO: stderr: "" +Feb 25 10:58:41.623: INFO: stdout: "update-demo-nautilus-jj9th update-demo-nautilus-xcrxn " +Feb 25 10:58:41.623: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-jj9th -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' +Feb 25 10:58:41.856: INFO: stderr: "" +Feb 25 10:58:41.856: INFO: stdout: "" +Feb 25 10:58:41.856: INFO: update-demo-nautilus-jj9th is created but not running +Feb 25 10:58:46.859: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' +Feb 25 10:58:47.082: INFO: stderr: "" +Feb 25 10:58:47.082: INFO: stdout: "update-demo-nautilus-jj9th update-demo-nautilus-xcrxn " +Feb 25 10:58:47.083: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-jj9th -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' +Feb 25 10:58:47.355: INFO: stderr: "" +Feb 25 10:58:47.355: INFO: stdout: "true" +Feb 25 10:58:47.355: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-jj9th -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' +Feb 25 10:58:47.537: INFO: stderr: "" +Feb 25 10:58:47.538: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" +Feb 25 10:58:47.538: INFO: validating pod update-demo-nautilus-jj9th +Feb 25 10:58:47.574: INFO: got data: { + "image": "nautilus.jpg" +} + +Feb 25 10:58:47.574: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Feb 25 10:58:47.575: INFO: update-demo-nautilus-jj9th is verified up and running +Feb 25 10:58:47.575: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-xcrxn -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' +Feb 25 10:58:47.803: INFO: stderr: "" +Feb 25 10:58:47.803: INFO: stdout: "true" +Feb 25 10:58:47.803: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-xcrxn -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' +Feb 25 10:58:48.008: INFO: stderr: "" +Feb 25 10:58:48.008: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" +Feb 25 10:58:48.008: INFO: validating pod update-demo-nautilus-xcrxn +Feb 25 10:58:48.023: INFO: got data: { + "image": "nautilus.jpg" +} + +Feb 25 10:58:48.023: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Feb 25 10:58:48.023: INFO: update-demo-nautilus-xcrxn is verified up and running +STEP: scaling down the replication controller 02/25/23 10:58:48.023 +Feb 25 10:58:48.052: INFO: scanned /root for discovery docs: +Feb 25 10:58:48.052: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 scale rc update-demo-nautilus --replicas=1 --timeout=5m' +Feb 25 10:58:49.368: INFO: stderr: "" +Feb 25 10:58:49.368: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" +STEP: waiting for all containers in name=update-demo pods to come up. 02/25/23 10:58:49.368 +Feb 25 10:58:49.369: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' +Feb 25 10:58:49.604: INFO: stderr: "" +Feb 25 10:58:49.604: INFO: stdout: "update-demo-nautilus-xcrxn " +Feb 25 10:58:49.605: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-xcrxn -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' +Feb 25 10:58:49.946: INFO: stderr: "" +Feb 25 10:58:49.946: INFO: stdout: "true" +Feb 25 10:58:49.947: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-xcrxn -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' +Feb 25 10:58:50.129: INFO: stderr: "" +Feb 25 10:58:50.129: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" +Feb 25 10:58:50.129: INFO: validating pod update-demo-nautilus-xcrxn +Feb 25 10:58:50.142: INFO: got data: { + "image": "nautilus.jpg" +} + +Feb 25 10:58:50.142: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Feb 25 10:58:50.142: INFO: update-demo-nautilus-xcrxn is verified up and running +STEP: scaling up the replication controller 02/25/23 10:58:50.142 +Feb 25 10:58:50.163: INFO: scanned /root for discovery docs: +Feb 25 10:58:50.164: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 scale rc update-demo-nautilus --replicas=2 --timeout=5m' +Feb 25 10:58:51.456: INFO: stderr: "" +Feb 25 10:58:51.456: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" +STEP: waiting for all containers in name=update-demo pods to come up. 02/25/23 10:58:51.457 +Feb 25 10:58:51.458: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' +Feb 25 10:58:51.691: INFO: stderr: "" +Feb 25 10:58:51.691: INFO: stdout: "update-demo-nautilus-n5jqg update-demo-nautilus-xcrxn " +Feb 25 10:58:51.692: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-n5jqg -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' +Feb 25 10:58:52.019: INFO: stderr: "" +Feb 25 10:58:52.019: INFO: stdout: "" +Feb 25 10:58:52.019: INFO: update-demo-nautilus-n5jqg is created but not running +Feb 25 10:58:57.023: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' +Feb 25 10:58:57.233: INFO: stderr: "" +Feb 25 10:58:57.233: INFO: stdout: "update-demo-nautilus-n5jqg update-demo-nautilus-xcrxn " +Feb 25 10:58:57.233: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-n5jqg -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' +Feb 25 10:58:57.411: INFO: stderr: "" +Feb 25 10:58:57.411: INFO: stdout: "true" +Feb 25 10:58:57.411: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-n5jqg -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' +Feb 25 10:58:57.609: INFO: stderr: "" +Feb 25 10:58:57.609: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" +Feb 25 10:58:57.609: INFO: validating pod update-demo-nautilus-n5jqg +Feb 25 10:58:57.634: INFO: got data: { + "image": "nautilus.jpg" +} + +Feb 25 10:58:57.634: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Feb 25 10:58:57.634: INFO: update-demo-nautilus-n5jqg is verified up and running +Feb 25 10:58:57.636: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-xcrxn -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' +Feb 25 10:58:57.890: INFO: stderr: "" +Feb 25 10:58:57.890: INFO: stdout: "true" +Feb 25 10:58:57.891: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-xcrxn -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' +Feb 25 10:58:58.075: INFO: stderr: "" +Feb 25 10:58:58.075: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" +Feb 25 10:58:58.075: INFO: validating pod update-demo-nautilus-xcrxn +Feb 25 10:58:58.087: INFO: got data: { + "image": "nautilus.jpg" +} + +Feb 25 10:58:58.087: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . +Feb 25 10:58:58.087: INFO: update-demo-nautilus-xcrxn is verified up and running +STEP: using delete to clean up resources 02/25/23 10:58:58.087 +Feb 25 10:58:58.088: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 delete --grace-period=0 --force -f -' +Feb 25 10:58:58.270: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Feb 25 10:58:58.270: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" +Feb 25 10:58:58.271: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get rc,svc -l name=update-demo --no-headers' +Feb 25 10:58:58.668: INFO: stderr: "No resources found in kubectl-3605 namespace.\n" +Feb 25 10:58:58.669: INFO: stdout: "" +Feb 25 10:58:58.669: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods -l name=update-demo -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' +Feb 25 10:58:59.036: INFO: stderr: "" +Feb 25 10:58:59.036: INFO: stdout: "" [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 -Dec 14 16:41:43.754: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-2615" for this suite. 12/14/22 16:41:43.761 -{"msg":"PASSED [sig-cli] Kubectl client Kubectl replace should update a single-container pod's image [Conformance]","completed":220,"skipped":4089,"failed":0} +Feb 25 10:58:59.036: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-3605" for this suite. 02/25/23 10:58:59.084 +{"msg":"PASSED [sig-cli] Kubectl client Update Demo should scale a replication controller [Conformance]","completed":217,"skipped":3990,"failed":0} ------------------------------ -• [SLOW TEST] [7.835 seconds] +• [SLOW TEST] [18.794 seconds] [sig-cli] Kubectl client test/e2e/kubectl/framework.go:23 - Kubectl replace - test/e2e/kubectl/kubectl.go:1729 - should update a single-container pod's image [Conformance] - test/e2e/kubectl/kubectl.go:1745 + Update Demo + test/e2e/kubectl/kubectl.go:322 + should scale a replication controller [Conformance] + test/e2e/kubectl/kubectl.go:350 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:41:35.935 - Dec 14 16:41:35.935: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 16:41:35.939 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:35.965 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:35.969 + STEP: Creating a kubernetes client 02/25/23 10:58:40.321 + Feb 25 10:58:40.321: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 10:58:40.324 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:58:40.352 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:58:40.36 [BeforeEach] [sig-cli] Kubectl client test/e2e/kubectl/kubectl.go:272 - [BeforeEach] Kubectl replace - test/e2e/kubectl/kubectl.go:1732 - [It] should update a single-container pod's image [Conformance] - test/e2e/kubectl/kubectl.go:1745 - STEP: running the image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 12/14/22 16:41:35.972 - Dec 14 16:41:35.972: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2615 run e2e-test-httpd-pod --image=registry.k8s.io/e2e-test-images/httpd:2.4.38-2 --pod-running-timeout=2m0s --labels=run=e2e-test-httpd-pod' - Dec 14 16:41:36.155: INFO: stderr: "" - Dec 14 16:41:36.156: INFO: stdout: "pod/e2e-test-httpd-pod created\n" - STEP: verifying the pod e2e-test-httpd-pod is running 12/14/22 16:41:36.156 - STEP: verifying the pod e2e-test-httpd-pod was created 12/14/22 16:41:41.208 - Dec 14 16:41:41.209: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2615 get pod e2e-test-httpd-pod -o json' - Dec 14 16:41:41.337: INFO: stderr: "" - Dec 14 16:41:41.337: INFO: stdout: "{\n \"apiVersion\": \"v1\",\n \"kind\": \"Pod\",\n \"metadata\": {\n \"creationTimestamp\": \"2022-12-14T16:41:36Z\",\n \"labels\": {\n \"run\": \"e2e-test-httpd-pod\"\n },\n \"name\": \"e2e-test-httpd-pod\",\n \"namespace\": \"kubectl-2615\",\n \"resourceVersion\": \"22676\",\n \"uid\": \"b5c068db-1a16-4c2f-ab3a-2dc68c28e44e\"\n },\n \"spec\": {\n \"containers\": [\n {\n \"image\": \"registry.k8s.io/e2e-test-images/httpd:2.4.38-2\",\n \"imagePullPolicy\": \"IfNotPresent\",\n \"name\": \"e2e-test-httpd-pod\",\n \"resources\": {},\n \"terminationMessagePath\": \"/dev/termination-log\",\n \"terminationMessagePolicy\": \"File\",\n \"volumeMounts\": [\n {\n \"mountPath\": \"/var/run/secrets/kubernetes.io/serviceaccount\",\n \"name\": \"kube-api-access-nq7k2\",\n \"readOnly\": true\n }\n ]\n }\n ],\n \"dnsPolicy\": \"ClusterFirst\",\n \"enableServiceLinks\": true,\n \"nodeName\": \"kiedai4iakae-3\",\n \"preemptionPolicy\": \"PreemptLowerPriority\",\n \"priority\": 0,\n \"restartPolicy\": \"Always\",\n \"schedulerName\": \"default-scheduler\",\n \"securityContext\": {},\n \"serviceAccount\": \"default\",\n \"serviceAccountName\": \"default\",\n \"terminationGracePeriodSeconds\": 30,\n \"tolerations\": [\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/not-ready\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n },\n {\n \"effect\": \"NoExecute\",\n \"key\": \"node.kubernetes.io/unreachable\",\n \"operator\": \"Exists\",\n \"tolerationSeconds\": 300\n }\n ],\n \"volumes\": [\n {\n \"name\": \"kube-api-access-nq7k2\",\n \"projected\": {\n \"defaultMode\": 420,\n \"sources\": [\n {\n \"serviceAccountToken\": {\n \"expirationSeconds\": 3607,\n \"path\": \"token\"\n }\n },\n {\n \"configMap\": {\n \"items\": [\n {\n \"key\": \"ca.crt\",\n \"path\": \"ca.crt\"\n }\n ],\n \"name\": \"kube-root-ca.crt\"\n }\n },\n {\n \"downwardAPI\": {\n \"items\": [\n {\n \"fieldRef\": {\n \"apiVersion\": \"v1\",\n \"fieldPath\": \"metadata.namespace\"\n },\n \"path\": \"namespace\"\n }\n ]\n }\n }\n ]\n }\n }\n ]\n },\n \"status\": {\n \"conditions\": [\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2022-12-14T16:41:36Z\",\n \"status\": \"True\",\n \"type\": \"Initialized\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2022-12-14T16:41:37Z\",\n \"status\": \"True\",\n \"type\": \"Ready\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2022-12-14T16:41:37Z\",\n \"status\": \"True\",\n \"type\": \"ContainersReady\"\n },\n {\n \"lastProbeTime\": null,\n \"lastTransitionTime\": \"2022-12-14T16:41:36Z\",\n \"status\": \"True\",\n \"type\": \"PodScheduled\"\n }\n ],\n \"containerStatuses\": [\n {\n \"containerID\": \"cri-o://1d429c9067a346e37eb8a6cb447ed6132b96763a0ea07f1c8c42d040e5d3d23b\",\n \"image\": \"registry.k8s.io/e2e-test-images/httpd:2.4.38-2\",\n \"imageID\": \"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3\",\n \"lastState\": {},\n \"name\": \"e2e-test-httpd-pod\",\n \"ready\": true,\n \"restartCount\": 0,\n \"started\": true,\n \"state\": {\n \"running\": {\n \"startedAt\": \"2022-12-14T16:41:37Z\"\n }\n }\n }\n ],\n \"hostIP\": \"192.168.121.244\",\n \"phase\": \"Running\",\n \"podIP\": \"10.233.66.60\",\n \"podIPs\": [\n {\n \"ip\": \"10.233.66.60\"\n }\n ],\n \"qosClass\": \"BestEffort\",\n \"startTime\": \"2022-12-14T16:41:36Z\"\n }\n}\n" - STEP: replace the image in the pod 12/14/22 16:41:41.338 - Dec 14 16:41:41.338: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2615 replace -f -' - Dec 14 16:41:41.664: INFO: stderr: "" - Dec 14 16:41:41.664: INFO: stdout: "pod/e2e-test-httpd-pod replaced\n" - STEP: verifying the pod e2e-test-httpd-pod has the right image registry.k8s.io/e2e-test-images/busybox:1.29-2 12/14/22 16:41:41.665 - [AfterEach] Kubectl replace - test/e2e/kubectl/kubectl.go:1736 - Dec 14 16:41:41.694: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2615 delete pods e2e-test-httpd-pod' - Dec 14 16:41:43.753: INFO: stderr: "" - Dec 14 16:41:43.753: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" - [AfterEach] [sig-cli] Kubectl client - test/e2e/framework/framework.go:187 - Dec 14 16:41:43.754: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-2615" for this suite. 12/14/22 16:41:43.761 - << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-node] Lease - lease API should be available [Conformance] - test/e2e/common/node/lease.go:72 -[BeforeEach] [sig-node] Lease - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:41:43.786 -Dec 14 16:41:43.786: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename lease-test 12/14/22 16:41:43.788 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:43.827 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:43.829 -[It] lease API should be available [Conformance] - test/e2e/common/node/lease.go:72 -[AfterEach] [sig-node] Lease - test/e2e/framework/framework.go:187 -Dec 14 16:41:43.930: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "lease-test-4733" for this suite. 12/14/22 16:41:43.935 -{"msg":"PASSED [sig-node] Lease lease API should be available [Conformance]","completed":221,"skipped":4142,"failed":0} ------------------------------- -• [0.158 seconds] -[sig-node] Lease -test/e2e/common/node/framework.go:23 - lease API should be available [Conformance] - test/e2e/common/node/lease.go:72 + [BeforeEach] Update Demo + test/e2e/kubectl/kubectl.go:324 + [It] should scale a replication controller [Conformance] + test/e2e/kubectl/kubectl.go:350 + STEP: creating a replication controller 02/25/23 10:58:40.368 + Feb 25 10:58:40.369: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 create -f -' + Feb 25 10:58:41.333: INFO: stderr: "" + Feb 25 10:58:41.333: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" + STEP: waiting for all containers in name=update-demo pods to come up. 02/25/23 10:58:41.333 + Feb 25 10:58:41.335: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' + Feb 25 10:58:41.623: INFO: stderr: "" + Feb 25 10:58:41.623: INFO: stdout: "update-demo-nautilus-jj9th update-demo-nautilus-xcrxn " + Feb 25 10:58:41.623: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-jj9th -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' + Feb 25 10:58:41.856: INFO: stderr: "" + Feb 25 10:58:41.856: INFO: stdout: "" + Feb 25 10:58:41.856: INFO: update-demo-nautilus-jj9th is created but not running + Feb 25 10:58:46.859: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' + Feb 25 10:58:47.082: INFO: stderr: "" + Feb 25 10:58:47.082: INFO: stdout: "update-demo-nautilus-jj9th update-demo-nautilus-xcrxn " + Feb 25 10:58:47.083: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-jj9th -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' + Feb 25 10:58:47.355: INFO: stderr: "" + Feb 25 10:58:47.355: INFO: stdout: "true" + Feb 25 10:58:47.355: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-jj9th -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' + Feb 25 10:58:47.537: INFO: stderr: "" + Feb 25 10:58:47.538: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" + Feb 25 10:58:47.538: INFO: validating pod update-demo-nautilus-jj9th + Feb 25 10:58:47.574: INFO: got data: { + "image": "nautilus.jpg" + } - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Lease - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:41:43.786 - Dec 14 16:41:43.786: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename lease-test 12/14/22 16:41:43.788 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:43.827 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:43.829 - [It] lease API should be available [Conformance] - test/e2e/common/node/lease.go:72 - [AfterEach] [sig-node] Lease + Feb 25 10:58:47.574: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . + Feb 25 10:58:47.575: INFO: update-demo-nautilus-jj9th is verified up and running + Feb 25 10:58:47.575: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-xcrxn -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' + Feb 25 10:58:47.803: INFO: stderr: "" + Feb 25 10:58:47.803: INFO: stdout: "true" + Feb 25 10:58:47.803: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-xcrxn -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' + Feb 25 10:58:48.008: INFO: stderr: "" + Feb 25 10:58:48.008: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" + Feb 25 10:58:48.008: INFO: validating pod update-demo-nautilus-xcrxn + Feb 25 10:58:48.023: INFO: got data: { + "image": "nautilus.jpg" + } + + Feb 25 10:58:48.023: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . + Feb 25 10:58:48.023: INFO: update-demo-nautilus-xcrxn is verified up and running + STEP: scaling down the replication controller 02/25/23 10:58:48.023 + Feb 25 10:58:48.052: INFO: scanned /root for discovery docs: + Feb 25 10:58:48.052: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 scale rc update-demo-nautilus --replicas=1 --timeout=5m' + Feb 25 10:58:49.368: INFO: stderr: "" + Feb 25 10:58:49.368: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" + STEP: waiting for all containers in name=update-demo pods to come up. 02/25/23 10:58:49.368 + Feb 25 10:58:49.369: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' + Feb 25 10:58:49.604: INFO: stderr: "" + Feb 25 10:58:49.604: INFO: stdout: "update-demo-nautilus-xcrxn " + Feb 25 10:58:49.605: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-xcrxn -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' + Feb 25 10:58:49.946: INFO: stderr: "" + Feb 25 10:58:49.946: INFO: stdout: "true" + Feb 25 10:58:49.947: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-xcrxn -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' + Feb 25 10:58:50.129: INFO: stderr: "" + Feb 25 10:58:50.129: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" + Feb 25 10:58:50.129: INFO: validating pod update-demo-nautilus-xcrxn + Feb 25 10:58:50.142: INFO: got data: { + "image": "nautilus.jpg" + } + + Feb 25 10:58:50.142: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . + Feb 25 10:58:50.142: INFO: update-demo-nautilus-xcrxn is verified up and running + STEP: scaling up the replication controller 02/25/23 10:58:50.142 + Feb 25 10:58:50.163: INFO: scanned /root for discovery docs: + Feb 25 10:58:50.164: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 scale rc update-demo-nautilus --replicas=2 --timeout=5m' + Feb 25 10:58:51.456: INFO: stderr: "" + Feb 25 10:58:51.456: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" + STEP: waiting for all containers in name=update-demo pods to come up. 02/25/23 10:58:51.457 + Feb 25 10:58:51.458: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' + Feb 25 10:58:51.691: INFO: stderr: "" + Feb 25 10:58:51.691: INFO: stdout: "update-demo-nautilus-n5jqg update-demo-nautilus-xcrxn " + Feb 25 10:58:51.692: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-n5jqg -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' + Feb 25 10:58:52.019: INFO: stderr: "" + Feb 25 10:58:52.019: INFO: stdout: "" + Feb 25 10:58:52.019: INFO: update-demo-nautilus-n5jqg is created but not running + Feb 25 10:58:57.023: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' + Feb 25 10:58:57.233: INFO: stderr: "" + Feb 25 10:58:57.233: INFO: stdout: "update-demo-nautilus-n5jqg update-demo-nautilus-xcrxn " + Feb 25 10:58:57.233: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-n5jqg -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' + Feb 25 10:58:57.411: INFO: stderr: "" + Feb 25 10:58:57.411: INFO: stdout: "true" + Feb 25 10:58:57.411: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-n5jqg -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' + Feb 25 10:58:57.609: INFO: stderr: "" + Feb 25 10:58:57.609: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" + Feb 25 10:58:57.609: INFO: validating pod update-demo-nautilus-n5jqg + Feb 25 10:58:57.634: INFO: got data: { + "image": "nautilus.jpg" + } + + Feb 25 10:58:57.634: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . + Feb 25 10:58:57.634: INFO: update-demo-nautilus-n5jqg is verified up and running + Feb 25 10:58:57.636: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-xcrxn -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' + Feb 25 10:58:57.890: INFO: stderr: "" + Feb 25 10:58:57.890: INFO: stdout: "true" + Feb 25 10:58:57.891: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods update-demo-nautilus-xcrxn -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' + Feb 25 10:58:58.075: INFO: stderr: "" + Feb 25 10:58:58.075: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" + Feb 25 10:58:58.075: INFO: validating pod update-demo-nautilus-xcrxn + Feb 25 10:58:58.087: INFO: got data: { + "image": "nautilus.jpg" + } + + Feb 25 10:58:58.087: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . + Feb 25 10:58:58.087: INFO: update-demo-nautilus-xcrxn is verified up and running + STEP: using delete to clean up resources 02/25/23 10:58:58.087 + Feb 25 10:58:58.088: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 delete --grace-period=0 --force -f -' + Feb 25 10:58:58.270: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" + Feb 25 10:58:58.270: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" + Feb 25 10:58:58.271: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get rc,svc -l name=update-demo --no-headers' + Feb 25 10:58:58.668: INFO: stderr: "No resources found in kubectl-3605 namespace.\n" + Feb 25 10:58:58.669: INFO: stdout: "" + Feb 25 10:58:58.669: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3605 get pods -l name=update-demo -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' + Feb 25 10:58:59.036: INFO: stderr: "" + Feb 25 10:58:59.036: INFO: stdout: "" + [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 16:41:43.930: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "lease-test-4733" for this suite. 12/14/22 16:41:43.935 + Feb 25 10:58:59.036: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-3605" for this suite. 02/25/23 10:58:59.084 << End Captured GinkgoWriter Output ------------------------------ -SS +SSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Secrets - should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:46 -[BeforeEach] [sig-storage] Secrets +[sig-apps] Deployment + deployment should support rollover [Conformance] + test/e2e/apps/deployment.go:132 +[BeforeEach] [sig-apps] Deployment test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:41:43.944 -Dec 14 16:41:43.945: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename secrets 12/14/22 16:41:43.948 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:43.973 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:43.976 -[It] should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:46 -STEP: Creating secret with name secret-test-464e7518-cc72-41bc-b31c-0f11f05cf5b0 12/14/22 16:41:43.979 -STEP: Creating a pod to test consume secrets 12/14/22 16:41:43.984 -Dec 14 16:41:43.996: INFO: Waiting up to 5m0s for pod "pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b" in namespace "secrets-4282" to be "Succeeded or Failed" -Dec 14 16:41:44.001: INFO: Pod "pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.781403ms -Dec 14 16:41:46.008: INFO: Pod "pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012128033s -Dec 14 16:41:48.010: INFO: Pod "pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014147555s -STEP: Saw pod success 12/14/22 16:41:48.01 -Dec 14 16:41:48.010: INFO: Pod "pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b" satisfied condition "Succeeded or Failed" -Dec 14 16:41:48.015: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b container secret-volume-test: -STEP: delete the pod 12/14/22 16:41:48.148 -Dec 14 16:41:48.183: INFO: Waiting for pod pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b to disappear -Dec 14 16:41:48.187: INFO: Pod pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b no longer exists -[AfterEach] [sig-storage] Secrets +STEP: Creating a kubernetes client 02/25/23 10:58:59.118 +Feb 25 10:58:59.118: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename deployment 02/25/23 10:58:59.13 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:58:59.209 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:58:59.218 +[BeforeEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:91 +[It] deployment should support rollover [Conformance] + test/e2e/apps/deployment.go:132 +Feb 25 10:58:59.266: INFO: Pod name rollover-pod: Found 0 pods out of 1 +Feb 25 10:59:04.278: INFO: Pod name rollover-pod: Found 1 pods out of 1 +STEP: ensuring each pod is running 02/25/23 10:59:04.279 +Feb 25 10:59:04.279: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready +Feb 25 10:59:06.289: INFO: Creating deployment "test-rollover-deployment" +Feb 25 10:59:06.311: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations +Feb 25 10:59:08.338: INFO: Check revision of new replica set for deployment "test-rollover-deployment" +Feb 25 10:59:08.356: INFO: Ensure that both replica sets have 1 created replica +Feb 25 10:59:08.371: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update +Feb 25 10:59:08.395: INFO: Updating deployment test-rollover-deployment +Feb 25 10:59:08.395: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller +Feb 25 10:59:10.420: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2 +Feb 25 10:59:10.437: INFO: Make sure deployment "test-rollover-deployment" is complete +Feb 25 10:59:10.455: INFO: all replica sets need to contain the pod-template-hash label +Feb 25 10:59:10.456: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 8, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:59:12.471: INFO: all replica sets need to contain the pod-template-hash label +Feb 25 10:59:12.471: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 10, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:59:14.480: INFO: all replica sets need to contain the pod-template-hash label +Feb 25 10:59:14.480: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 10, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:59:16.484: INFO: all replica sets need to contain the pod-template-hash label +Feb 25 10:59:16.484: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 10, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:59:18.478: INFO: all replica sets need to contain the pod-template-hash label +Feb 25 10:59:18.478: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 10, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:59:20.477: INFO: all replica sets need to contain the pod-template-hash label +Feb 25 10:59:20.477: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 10, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} +Feb 25 10:59:22.479: INFO: +Feb 25 10:59:22.480: INFO: Ensure that both old replica sets have no replicas +[AfterEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:84 +Feb 25 10:59:22.510: INFO: Deployment "test-rollover-deployment": +&Deployment{ObjectMeta:{test-rollover-deployment deployment-2755 a50991db-b886-4db5-a54f-0dec1e714166 33196 2 2023-02-25 10:59:06 +0000 UTC map[name:rollover-pod] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2023-02-25 10:59:08 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:minReadySeconds":{},"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:59:20 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0047f47a8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2023-02-25 10:59:06 +0000 UTC,LastTransitionTime:2023-02-25 10:59:06 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rollover-deployment-6d45fd857b" has successfully progressed.,LastUpdateTime:2023-02-25 10:59:20 +0000 UTC,LastTransitionTime:2023-02-25 10:59:06 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} + +Feb 25 10:59:22.519: INFO: New ReplicaSet "test-rollover-deployment-6d45fd857b" of Deployment "test-rollover-deployment": +&ReplicaSet{ObjectMeta:{test-rollover-deployment-6d45fd857b deployment-2755 c68b9af8-ac7e-4fbf-adf8-564d553aec7e 33185 2 2023-02-25 10:59:08 +0000 UTC map[name:rollover-pod pod-template-hash:6d45fd857b] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-rollover-deployment a50991db-b886-4db5-a54f-0dec1e714166 0xc0047f4d67 0xc0047f4d68}] [] [{kube-controller-manager Update apps/v1 2023-02-25 10:59:08 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"a50991db-b886-4db5-a54f-0dec1e714166\"}":{}}},"f:spec":{"f:minReadySeconds":{},"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:59:20 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 6d45fd857b,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod-template-hash:6d45fd857b] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0047f4e18 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} +Feb 25 10:59:22.519: INFO: All old ReplicaSets of Deployment "test-rollover-deployment": +Feb 25 10:59:22.519: INFO: &ReplicaSet{ObjectMeta:{test-rollover-controller deployment-2755 64dda0a0-58bd-40a1-9ff0-718f1614847c 33194 2 2023-02-25 10:58:59 +0000 UTC map[name:rollover-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2] [{apps/v1 Deployment test-rollover-deployment a50991db-b886-4db5-a54f-0dec1e714166 0xc0047f4b17 0xc0047f4b18}] [] [{e2e.test Update apps/v1 2023-02-25 10:58:59 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:59:20 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"a50991db-b886-4db5-a54f-0dec1e714166\"}":{}}},"f:spec":{"f:replicas":{}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:59:20 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc0047f4bd8 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} +Feb 25 10:59:22.519: INFO: &ReplicaSet{ObjectMeta:{test-rollover-deployment-59b9df946d deployment-2755 35028286-32cf-4fee-9234-a69c6e50c8a1 33143 2 2023-02-25 10:59:06 +0000 UTC map[name:rollover-pod pod-template-hash:59b9df946d] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-rollover-deployment a50991db-b886-4db5-a54f-0dec1e714166 0xc0047f4c47 0xc0047f4c48}] [] [{kube-controller-manager Update apps/v1 2023-02-25 10:59:08 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"a50991db-b886-4db5-a54f-0dec1e714166\"}":{}}},"f:spec":{"f:minReadySeconds":{},"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"redis-slave\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:59:08 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 59b9df946d,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod-template-hash:59b9df946d] map[] [] [] []} {[] [] [{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0047f4cf8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} +Feb 25 10:59:22.529: INFO: Pod "test-rollover-deployment-6d45fd857b-6xkgm" is available: +&Pod{ObjectMeta:{test-rollover-deployment-6d45fd857b-6xkgm test-rollover-deployment-6d45fd857b- deployment-2755 9270d693-2e4b-4ef2-93b2-5d9985fdfa84 33162 0 2023-02-25 10:59:08 +0000 UTC map[name:rollover-pod pod-template-hash:6d45fd857b] map[] [{apps/v1 ReplicaSet test-rollover-deployment-6d45fd857b c68b9af8-ac7e-4fbf-adf8-564d553aec7e 0xc0047f5357 0xc0047f5358}] [] [{kube-controller-manager Update v1 2023-02-25 10:59:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c68b9af8-ac7e-4fbf-adf8-564d553aec7e\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 10:59:10 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.32\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-j67df,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j67df,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:59:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:59:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:59:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:59:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:10.233.66.32,StartTime:2023-02-25 10:59:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 10:59:10 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,ImageID:registry.k8s.io/e2e-test-images/agnhost@sha256:37dd9ad84b34913b4bc57e00b0a28e8b00f18ca6a5ecec69461aa57402e5c787,ContainerID:cri-o://6375addbf696ab8b80764e61226d2845d17057ff3cbb5106312f349f44ddea81,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.32,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +[AfterEach] [sig-apps] Deployment test/e2e/framework/framework.go:187 -Dec 14 16:41:48.187: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-4282" for this suite. 12/14/22 16:41:48.193 -{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume [NodeConformance] [Conformance]","completed":222,"skipped":4144,"failed":0} +Feb 25 10:59:22.529: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "deployment-2755" for this suite. 02/25/23 10:59:22.547 +{"msg":"PASSED [sig-apps] Deployment deployment should support rollover [Conformance]","completed":218,"skipped":4011,"failed":0} ------------------------------ -• [4.260 seconds] -[sig-storage] Secrets -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:46 +• [SLOW TEST] [23.450 seconds] +[sig-apps] Deployment +test/e2e/apps/framework.go:23 + deployment should support rollover [Conformance] + test/e2e/apps/deployment.go:132 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Secrets + [BeforeEach] [sig-apps] Deployment test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:41:43.944 - Dec 14 16:41:43.945: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename secrets 12/14/22 16:41:43.948 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:43.973 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:43.976 - [It] should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:46 - STEP: Creating secret with name secret-test-464e7518-cc72-41bc-b31c-0f11f05cf5b0 12/14/22 16:41:43.979 - STEP: Creating a pod to test consume secrets 12/14/22 16:41:43.984 - Dec 14 16:41:43.996: INFO: Waiting up to 5m0s for pod "pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b" in namespace "secrets-4282" to be "Succeeded or Failed" - Dec 14 16:41:44.001: INFO: Pod "pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.781403ms - Dec 14 16:41:46.008: INFO: Pod "pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012128033s - Dec 14 16:41:48.010: INFO: Pod "pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014147555s - STEP: Saw pod success 12/14/22 16:41:48.01 - Dec 14 16:41:48.010: INFO: Pod "pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b" satisfied condition "Succeeded or Failed" - Dec 14 16:41:48.015: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b container secret-volume-test: - STEP: delete the pod 12/14/22 16:41:48.148 - Dec 14 16:41:48.183: INFO: Waiting for pod pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b to disappear - Dec 14 16:41:48.187: INFO: Pod pod-secrets-ca44e071-0bd6-4d09-9fb3-3e2083dcd70b no longer exists - [AfterEach] [sig-storage] Secrets + STEP: Creating a kubernetes client 02/25/23 10:58:59.118 + Feb 25 10:58:59.118: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename deployment 02/25/23 10:58:59.13 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:58:59.209 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:58:59.218 + [BeforeEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:91 + [It] deployment should support rollover [Conformance] + test/e2e/apps/deployment.go:132 + Feb 25 10:58:59.266: INFO: Pod name rollover-pod: Found 0 pods out of 1 + Feb 25 10:59:04.278: INFO: Pod name rollover-pod: Found 1 pods out of 1 + STEP: ensuring each pod is running 02/25/23 10:59:04.279 + Feb 25 10:59:04.279: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready + Feb 25 10:59:06.289: INFO: Creating deployment "test-rollover-deployment" + Feb 25 10:59:06.311: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations + Feb 25 10:59:08.338: INFO: Check revision of new replica set for deployment "test-rollover-deployment" + Feb 25 10:59:08.356: INFO: Ensure that both replica sets have 1 created replica + Feb 25 10:59:08.371: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update + Feb 25 10:59:08.395: INFO: Updating deployment test-rollover-deployment + Feb 25 10:59:08.395: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller + Feb 25 10:59:10.420: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2 + Feb 25 10:59:10.437: INFO: Make sure deployment "test-rollover-deployment" is complete + Feb 25 10:59:10.455: INFO: all replica sets need to contain the pod-template-hash label + Feb 25 10:59:10.456: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:1, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 8, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:59:12.471: INFO: all replica sets need to contain the pod-template-hash label + Feb 25 10:59:12.471: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 10, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:59:14.480: INFO: all replica sets need to contain the pod-template-hash label + Feb 25 10:59:14.480: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 10, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:59:16.484: INFO: all replica sets need to contain the pod-template-hash label + Feb 25 10:59:16.484: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 10, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:59:18.478: INFO: all replica sets need to contain the pod-template-hash label + Feb 25 10:59:18.478: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 10, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:59:20.477: INFO: all replica sets need to contain the pod-template-hash label + Feb 25 10:59:20.477: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 10, 59, 10, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 10, 59, 6, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} + Feb 25 10:59:22.479: INFO: + Feb 25 10:59:22.480: INFO: Ensure that both old replica sets have no replicas + [AfterEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:84 + Feb 25 10:59:22.510: INFO: Deployment "test-rollover-deployment": + &Deployment{ObjectMeta:{test-rollover-deployment deployment-2755 a50991db-b886-4db5-a54f-0dec1e714166 33196 2 2023-02-25 10:59:06 +0000 UTC map[name:rollover-pod] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2023-02-25 10:59:08 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:minReadySeconds":{},"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:59:20 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0047f47a8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2023-02-25 10:59:06 +0000 UTC,LastTransitionTime:2023-02-25 10:59:06 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rollover-deployment-6d45fd857b" has successfully progressed.,LastUpdateTime:2023-02-25 10:59:20 +0000 UTC,LastTransitionTime:2023-02-25 10:59:06 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} + + Feb 25 10:59:22.519: INFO: New ReplicaSet "test-rollover-deployment-6d45fd857b" of Deployment "test-rollover-deployment": + &ReplicaSet{ObjectMeta:{test-rollover-deployment-6d45fd857b deployment-2755 c68b9af8-ac7e-4fbf-adf8-564d553aec7e 33185 2 2023-02-25 10:59:08 +0000 UTC map[name:rollover-pod pod-template-hash:6d45fd857b] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-rollover-deployment a50991db-b886-4db5-a54f-0dec1e714166 0xc0047f4d67 0xc0047f4d68}] [] [{kube-controller-manager Update apps/v1 2023-02-25 10:59:08 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"a50991db-b886-4db5-a54f-0dec1e714166\"}":{}}},"f:spec":{"f:minReadySeconds":{},"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:59:20 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 6d45fd857b,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod-template-hash:6d45fd857b] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0047f4e18 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} + Feb 25 10:59:22.519: INFO: All old ReplicaSets of Deployment "test-rollover-deployment": + Feb 25 10:59:22.519: INFO: &ReplicaSet{ObjectMeta:{test-rollover-controller deployment-2755 64dda0a0-58bd-40a1-9ff0-718f1614847c 33194 2 2023-02-25 10:58:59 +0000 UTC map[name:rollover-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2] [{apps/v1 Deployment test-rollover-deployment a50991db-b886-4db5-a54f-0dec1e714166 0xc0047f4b17 0xc0047f4b18}] [] [{e2e.test Update apps/v1 2023-02-25 10:58:59 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:59:20 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"a50991db-b886-4db5-a54f-0dec1e714166\"}":{}}},"f:spec":{"f:replicas":{}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:59:20 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc0047f4bd8 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} + Feb 25 10:59:22.519: INFO: &ReplicaSet{ObjectMeta:{test-rollover-deployment-59b9df946d deployment-2755 35028286-32cf-4fee-9234-a69c6e50c8a1 33143 2 2023-02-25 10:59:06 +0000 UTC map[name:rollover-pod pod-template-hash:59b9df946d] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-rollover-deployment a50991db-b886-4db5-a54f-0dec1e714166 0xc0047f4c47 0xc0047f4c48}] [] [{kube-controller-manager Update apps/v1 2023-02-25 10:59:08 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"a50991db-b886-4db5-a54f-0dec1e714166\"}":{}}},"f:spec":{"f:minReadySeconds":{},"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"redis-slave\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 10:59:08 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 59b9df946d,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod-template-hash:59b9df946d] map[] [] [] []} {[] [] [{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0047f4cf8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} + Feb 25 10:59:22.529: INFO: Pod "test-rollover-deployment-6d45fd857b-6xkgm" is available: + &Pod{ObjectMeta:{test-rollover-deployment-6d45fd857b-6xkgm test-rollover-deployment-6d45fd857b- deployment-2755 9270d693-2e4b-4ef2-93b2-5d9985fdfa84 33162 0 2023-02-25 10:59:08 +0000 UTC map[name:rollover-pod pod-template-hash:6d45fd857b] map[] [{apps/v1 ReplicaSet test-rollover-deployment-6d45fd857b c68b9af8-ac7e-4fbf-adf8-564d553aec7e 0xc0047f5357 0xc0047f5358}] [] [{kube-controller-manager Update v1 2023-02-25 10:59:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"c68b9af8-ac7e-4fbf-adf8-564d553aec7e\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 10:59:10 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.32\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-j67df,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j67df,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:59:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:59:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:59:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 10:59:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:10.233.66.32,StartTime:2023-02-25 10:59:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 10:59:10 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,ImageID:registry.k8s.io/e2e-test-images/agnhost@sha256:37dd9ad84b34913b4bc57e00b0a28e8b00f18ca6a5ecec69461aa57402e5c787,ContainerID:cri-o://6375addbf696ab8b80764e61226d2845d17057ff3cbb5106312f349f44ddea81,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.32,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + [AfterEach] [sig-apps] Deployment test/e2e/framework/framework.go:187 - Dec 14 16:41:48.187: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "secrets-4282" for this suite. 12/14/22 16:41:48.193 + Feb 25 10:59:22.529: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "deployment-2755" for this suite. 02/25/23 10:59:22.547 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSS +SSSSSSSSS ------------------------------ -[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] - should include custom resource definition resources in discovery documents [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:198 -[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +[sig-storage] ConfigMap + should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:98 +[BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:41:48.209 -Dec 14 16:41:48.209: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename custom-resource-definition 12/14/22 16:41:48.211 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:48.238 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:48.241 -[It] should include custom resource definition resources in discovery documents [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:198 -STEP: fetching the /apis discovery document 12/14/22 16:41:48.245 -STEP: finding the apiextensions.k8s.io API group in the /apis discovery document 12/14/22 16:41:48.247 -STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis discovery document 12/14/22 16:41:48.247 -STEP: fetching the /apis/apiextensions.k8s.io discovery document 12/14/22 16:41:48.248 -STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis/apiextensions.k8s.io discovery document 12/14/22 16:41:48.249 -STEP: fetching the /apis/apiextensions.k8s.io/v1 discovery document 12/14/22 16:41:48.249 -STEP: finding customresourcedefinitions resources in the /apis/apiextensions.k8s.io/v1 discovery document 12/14/22 16:41:48.253 -[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 10:59:22.573 +Feb 25 10:59:22.573: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename configmap 02/25/23 10:59:22.58 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:59:22.611 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:59:22.616 +[It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:98 +STEP: Creating configMap with name configmap-test-volume-map-f085dd92-2b49-4ca5-b6f9-c0be53834b02 02/25/23 10:59:22.623 +STEP: Creating a pod to test consume configMaps 02/25/23 10:59:22.638 +Feb 25 10:59:22.669: INFO: Waiting up to 5m0s for pod "pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659" in namespace "configmap-1194" to be "Succeeded or Failed" +Feb 25 10:59:22.677: INFO: Pod "pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659": Phase="Pending", Reason="", readiness=false. Elapsed: 8.150138ms +Feb 25 10:59:24.687: INFO: Pod "pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018058831s +Feb 25 10:59:26.688: INFO: Pod "pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659": Phase="Pending", Reason="", readiness=false. Elapsed: 4.018567654s +Feb 25 10:59:28.691: INFO: Pod "pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.02151468s +STEP: Saw pod success 02/25/23 10:59:28.691 +Feb 25 10:59:28.692: INFO: Pod "pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659" satisfied condition "Succeeded or Failed" +Feb 25 10:59:28.703: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659 container agnhost-container: +STEP: delete the pod 02/25/23 10:59:28.75 +Feb 25 10:59:28.783: INFO: Waiting for pod pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659 to disappear +Feb 25 10:59:28.793: INFO: Pod pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659 no longer exists +[AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 -Dec 14 16:41:48.253: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "custom-resource-definition-4402" for this suite. 12/14/22 16:41:48.266 -{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] should include custom resource definition resources in discovery documents [Conformance]","completed":223,"skipped":4155,"failed":0} +Feb 25 10:59:28.794: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-1194" for this suite. 02/25/23 10:59:28.807 +{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance]","completed":219,"skipped":4020,"failed":0} ------------------------------ -• [0.068 seconds] -[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - should include custom resource definition resources in discovery documents [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:198 +• [SLOW TEST] [6.269 seconds] +[sig-storage] ConfigMap +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:98 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] + [BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:41:48.209 - Dec 14 16:41:48.209: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename custom-resource-definition 12/14/22 16:41:48.211 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:48.238 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:48.241 - [It] should include custom resource definition resources in discovery documents [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:198 - STEP: fetching the /apis discovery document 12/14/22 16:41:48.245 - STEP: finding the apiextensions.k8s.io API group in the /apis discovery document 12/14/22 16:41:48.247 - STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis discovery document 12/14/22 16:41:48.247 - STEP: fetching the /apis/apiextensions.k8s.io discovery document 12/14/22 16:41:48.248 - STEP: finding the apiextensions.k8s.io/v1 API group/version in the /apis/apiextensions.k8s.io discovery document 12/14/22 16:41:48.249 - STEP: fetching the /apis/apiextensions.k8s.io/v1 discovery document 12/14/22 16:41:48.249 - STEP: finding customresourcedefinitions resources in the /apis/apiextensions.k8s.io/v1 discovery document 12/14/22 16:41:48.253 - [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 10:59:22.573 + Feb 25 10:59:22.573: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename configmap 02/25/23 10:59:22.58 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:59:22.611 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:59:22.616 + [It] should be consumable from pods in volume with mappings and Item mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:98 + STEP: Creating configMap with name configmap-test-volume-map-f085dd92-2b49-4ca5-b6f9-c0be53834b02 02/25/23 10:59:22.623 + STEP: Creating a pod to test consume configMaps 02/25/23 10:59:22.638 + Feb 25 10:59:22.669: INFO: Waiting up to 5m0s for pod "pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659" in namespace "configmap-1194" to be "Succeeded or Failed" + Feb 25 10:59:22.677: INFO: Pod "pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659": Phase="Pending", Reason="", readiness=false. Elapsed: 8.150138ms + Feb 25 10:59:24.687: INFO: Pod "pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018058831s + Feb 25 10:59:26.688: INFO: Pod "pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659": Phase="Pending", Reason="", readiness=false. Elapsed: 4.018567654s + Feb 25 10:59:28.691: INFO: Pod "pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.02151468s + STEP: Saw pod success 02/25/23 10:59:28.691 + Feb 25 10:59:28.692: INFO: Pod "pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659" satisfied condition "Succeeded or Failed" + Feb 25 10:59:28.703: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659 container agnhost-container: + STEP: delete the pod 02/25/23 10:59:28.75 + Feb 25 10:59:28.783: INFO: Waiting for pod pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659 to disappear + Feb 25 10:59:28.793: INFO: Pod pod-configmaps-5ac657ba-e778-4b9d-b232-486fa2702659 no longer exists + [AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 - Dec 14 16:41:48.253: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "custom-resource-definition-4402" for this suite. 12/14/22 16:41:48.266 + Feb 25 10:59:28.794: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "configmap-1194" for this suite. 02/25/23 10:59:28.807 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Security Context - should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] - test/e2e/node/security_context.go:132 -[BeforeEach] [sig-node] Security Context +[sig-storage] Projected secret + should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:118 +[BeforeEach] [sig-storage] Projected secret test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:41:48.277 -Dec 14 16:41:48.278: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename security-context 12/14/22 16:41:48.282 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:48.302 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:48.305 -[It] should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] - test/e2e/node/security_context.go:132 -STEP: Creating a pod to test pod.Spec.SecurityContext.RunAsUser 12/14/22 16:41:48.309 -Dec 14 16:41:48.324: INFO: Waiting up to 5m0s for pod "security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c" in namespace "security-context-3095" to be "Succeeded or Failed" -Dec 14 16:41:48.333: INFO: Pod "security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c": Phase="Pending", Reason="", readiness=false. Elapsed: 8.803853ms -Dec 14 16:41:50.342: INFO: Pod "security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018044188s -Dec 14 16:41:52.343: INFO: Pod "security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018914489s -STEP: Saw pod success 12/14/22 16:41:52.343 -Dec 14 16:41:52.344: INFO: Pod "security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c" satisfied condition "Succeeded or Failed" -Dec 14 16:41:52.350: INFO: Trying to get logs from node kiedai4iakae-3 pod security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c container test-container: -STEP: delete the pod 12/14/22 16:41:52.362 -Dec 14 16:41:52.385: INFO: Waiting for pod security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c to disappear -Dec 14 16:41:52.391: INFO: Pod security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c no longer exists -[AfterEach] [sig-node] Security Context +STEP: Creating a kubernetes client 02/25/23 10:59:28.845 +Feb 25 10:59:28.845: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 10:59:28.851 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:59:28.921 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:59:28.946 +[It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:118 +STEP: Creating secret with name projected-secret-test-f1881105-ea44-4b7c-8862-f8a645129519 02/25/23 10:59:28.963 +STEP: Creating a pod to test consume secrets 02/25/23 10:59:28.98 +Feb 25 10:59:29.057: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304" in namespace "projected-7422" to be "Succeeded or Failed" +Feb 25 10:59:29.143: INFO: Pod "pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304": Phase="Pending", Reason="", readiness=false. Elapsed: 85.182869ms +Feb 25 10:59:31.151: INFO: Pod "pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304": Phase="Pending", Reason="", readiness=false. Elapsed: 2.093200913s +Feb 25 10:59:33.300: INFO: Pod "pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304": Phase="Pending", Reason="", readiness=false. Elapsed: 4.242507428s +Feb 25 10:59:35.151: INFO: Pod "pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.093670169s +STEP: Saw pod success 02/25/23 10:59:35.151 +Feb 25 10:59:35.152: INFO: Pod "pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304" satisfied condition "Succeeded or Failed" +Feb 25 10:59:35.176: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304 container secret-volume-test: +STEP: delete the pod 02/25/23 10:59:35.211 +Feb 25 10:59:35.251: INFO: Waiting for pod pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304 to disappear +Feb 25 10:59:35.257: INFO: Pod pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304 no longer exists +[AfterEach] [sig-storage] Projected secret test/e2e/framework/framework.go:187 -Dec 14 16:41:52.391: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "security-context-3095" for this suite. 12/14/22 16:41:52.398 -{"msg":"PASSED [sig-node] Security Context should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance]","completed":224,"skipped":4167,"failed":0} +Feb 25 10:59:35.258: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-7422" for this suite. 02/25/23 10:59:35.281 +{"msg":"PASSED [sig-storage] Projected secret should be consumable in multiple volumes in a pod [NodeConformance] [Conformance]","completed":220,"skipped":4045,"failed":0} ------------------------------ -• [4.131 seconds] -[sig-node] Security Context -test/e2e/node/framework.go:23 - should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] - test/e2e/node/security_context.go:132 +• [SLOW TEST] [6.452 seconds] +[sig-storage] Projected secret +test/e2e/common/storage/framework.go:23 + should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:118 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Security Context + [BeforeEach] [sig-storage] Projected secret test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:41:48.277 - Dec 14 16:41:48.278: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename security-context 12/14/22 16:41:48.282 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:48.302 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:48.305 - [It] should support container.SecurityContext.RunAsUser And container.SecurityContext.RunAsGroup [LinuxOnly] [Conformance] - test/e2e/node/security_context.go:132 - STEP: Creating a pod to test pod.Spec.SecurityContext.RunAsUser 12/14/22 16:41:48.309 - Dec 14 16:41:48.324: INFO: Waiting up to 5m0s for pod "security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c" in namespace "security-context-3095" to be "Succeeded or Failed" - Dec 14 16:41:48.333: INFO: Pod "security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c": Phase="Pending", Reason="", readiness=false. Elapsed: 8.803853ms - Dec 14 16:41:50.342: INFO: Pod "security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018044188s - Dec 14 16:41:52.343: INFO: Pod "security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018914489s - STEP: Saw pod success 12/14/22 16:41:52.343 - Dec 14 16:41:52.344: INFO: Pod "security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c" satisfied condition "Succeeded or Failed" - Dec 14 16:41:52.350: INFO: Trying to get logs from node kiedai4iakae-3 pod security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c container test-container: - STEP: delete the pod 12/14/22 16:41:52.362 - Dec 14 16:41:52.385: INFO: Waiting for pod security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c to disappear - Dec 14 16:41:52.391: INFO: Pod security-context-7ef3ef96-9af8-4365-b4d5-4765633c1e7c no longer exists - [AfterEach] [sig-node] Security Context + STEP: Creating a kubernetes client 02/25/23 10:59:28.845 + Feb 25 10:59:28.845: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 10:59:28.851 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:59:28.921 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:59:28.946 + [It] should be consumable in multiple volumes in a pod [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:118 + STEP: Creating secret with name projected-secret-test-f1881105-ea44-4b7c-8862-f8a645129519 02/25/23 10:59:28.963 + STEP: Creating a pod to test consume secrets 02/25/23 10:59:28.98 + Feb 25 10:59:29.057: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304" in namespace "projected-7422" to be "Succeeded or Failed" + Feb 25 10:59:29.143: INFO: Pod "pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304": Phase="Pending", Reason="", readiness=false. Elapsed: 85.182869ms + Feb 25 10:59:31.151: INFO: Pod "pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304": Phase="Pending", Reason="", readiness=false. Elapsed: 2.093200913s + Feb 25 10:59:33.300: INFO: Pod "pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304": Phase="Pending", Reason="", readiness=false. Elapsed: 4.242507428s + Feb 25 10:59:35.151: INFO: Pod "pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.093670169s + STEP: Saw pod success 02/25/23 10:59:35.151 + Feb 25 10:59:35.152: INFO: Pod "pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304" satisfied condition "Succeeded or Failed" + Feb 25 10:59:35.176: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304 container secret-volume-test: + STEP: delete the pod 02/25/23 10:59:35.211 + Feb 25 10:59:35.251: INFO: Waiting for pod pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304 to disappear + Feb 25 10:59:35.257: INFO: Pod pod-projected-secrets-38fdfa40-1928-4efe-b972-1760cf414304 no longer exists + [AfterEach] [sig-storage] Projected secret test/e2e/framework/framework.go:187 - Dec 14 16:41:52.391: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "security-context-3095" for this suite. 12/14/22 16:41:52.398 + Feb 25 10:59:35.258: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-7422" for this suite. 02/25/23 10:59:35.281 << End Captured GinkgoWriter Output ------------------------------ -SS +SSSSSSSSS ------------------------------ -[sig-api-machinery] ResourceQuota - should verify ResourceQuota with best effort scope. [Conformance] - test/e2e/apimachinery/resource_quota.go:793 -[BeforeEach] [sig-api-machinery] ResourceQuota +[sig-instrumentation] Events API + should delete a collection of events [Conformance] + test/e2e/instrumentation/events.go:207 +[BeforeEach] [sig-instrumentation] Events API test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:41:52.409 -Dec 14 16:41:52.409: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename resourcequota 12/14/22 16:41:52.411 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:52.443 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:52.447 -[It] should verify ResourceQuota with best effort scope. [Conformance] - test/e2e/apimachinery/resource_quota.go:793 -STEP: Creating a ResourceQuota with best effort scope 12/14/22 16:41:52.45 -STEP: Ensuring ResourceQuota status is calculated 12/14/22 16:41:52.459 -STEP: Creating a ResourceQuota with not best effort scope 12/14/22 16:41:54.465 -STEP: Ensuring ResourceQuota status is calculated 12/14/22 16:41:54.473 -STEP: Creating a best-effort pod 12/14/22 16:41:56.479 -STEP: Ensuring resource quota with best effort scope captures the pod usage 12/14/22 16:41:56.498 -STEP: Ensuring resource quota with not best effort ignored the pod usage 12/14/22 16:41:58.507 -STEP: Deleting the pod 12/14/22 16:42:00.514 -STEP: Ensuring resource quota status released the pod usage 12/14/22 16:42:00.532 -STEP: Creating a not best-effort pod 12/14/22 16:42:02.537 -STEP: Ensuring resource quota with not best effort scope captures the pod usage 12/14/22 16:42:02.554 -STEP: Ensuring resource quota with best effort scope ignored the pod usage 12/14/22 16:42:04.559 -STEP: Deleting the pod 12/14/22 16:42:06.566 -STEP: Ensuring resource quota status released the pod usage 12/14/22 16:42:06.585 -[AfterEach] [sig-api-machinery] ResourceQuota +STEP: Creating a kubernetes client 02/25/23 10:59:35.307 +Feb 25 10:59:35.308: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename events 02/25/23 10:59:35.31 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:59:35.341 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:59:35.383 +[BeforeEach] [sig-instrumentation] Events API + test/e2e/instrumentation/events.go:84 +[It] should delete a collection of events [Conformance] + test/e2e/instrumentation/events.go:207 +STEP: Create set of events 02/25/23 10:59:35.403 +STEP: get a list of Events with a label in the current namespace 02/25/23 10:59:35.482 +STEP: delete a list of events 02/25/23 10:59:35.507 +Feb 25 10:59:35.507: INFO: requesting DeleteCollection of events +STEP: check that the list of events matches the requested quantity 02/25/23 10:59:35.589 +[AfterEach] [sig-instrumentation] Events API test/e2e/framework/framework.go:187 -Dec 14 16:42:08.592: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-7588" for this suite. 12/14/22 16:42:08.598 -{"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope. [Conformance]","completed":225,"skipped":4169,"failed":0} +Feb 25 10:59:35.597: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "events-4930" for this suite. 02/25/23 10:59:35.628 +{"msg":"PASSED [sig-instrumentation] Events API should delete a collection of events [Conformance]","completed":221,"skipped":4054,"failed":0} ------------------------------ -• [SLOW TEST] [16.198 seconds] -[sig-api-machinery] ResourceQuota -test/e2e/apimachinery/framework.go:23 - should verify ResourceQuota with best effort scope. [Conformance] - test/e2e/apimachinery/resource_quota.go:793 +• [0.381 seconds] +[sig-instrumentation] Events API +test/e2e/instrumentation/common/framework.go:23 + should delete a collection of events [Conformance] + test/e2e/instrumentation/events.go:207 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] ResourceQuota + [BeforeEach] [sig-instrumentation] Events API test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:41:52.409 - Dec 14 16:41:52.409: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename resourcequota 12/14/22 16:41:52.411 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:41:52.443 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:41:52.447 - [It] should verify ResourceQuota with best effort scope. [Conformance] - test/e2e/apimachinery/resource_quota.go:793 - STEP: Creating a ResourceQuota with best effort scope 12/14/22 16:41:52.45 - STEP: Ensuring ResourceQuota status is calculated 12/14/22 16:41:52.459 - STEP: Creating a ResourceQuota with not best effort scope 12/14/22 16:41:54.465 - STEP: Ensuring ResourceQuota status is calculated 12/14/22 16:41:54.473 - STEP: Creating a best-effort pod 12/14/22 16:41:56.479 - STEP: Ensuring resource quota with best effort scope captures the pod usage 12/14/22 16:41:56.498 - STEP: Ensuring resource quota with not best effort ignored the pod usage 12/14/22 16:41:58.507 - STEP: Deleting the pod 12/14/22 16:42:00.514 - STEP: Ensuring resource quota status released the pod usage 12/14/22 16:42:00.532 - STEP: Creating a not best-effort pod 12/14/22 16:42:02.537 - STEP: Ensuring resource quota with not best effort scope captures the pod usage 12/14/22 16:42:02.554 - STEP: Ensuring resource quota with best effort scope ignored the pod usage 12/14/22 16:42:04.559 - STEP: Deleting the pod 12/14/22 16:42:06.566 - STEP: Ensuring resource quota status released the pod usage 12/14/22 16:42:06.585 - [AfterEach] [sig-api-machinery] ResourceQuota + STEP: Creating a kubernetes client 02/25/23 10:59:35.307 + Feb 25 10:59:35.308: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename events 02/25/23 10:59:35.31 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:59:35.341 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:59:35.383 + [BeforeEach] [sig-instrumentation] Events API + test/e2e/instrumentation/events.go:84 + [It] should delete a collection of events [Conformance] + test/e2e/instrumentation/events.go:207 + STEP: Create set of events 02/25/23 10:59:35.403 + STEP: get a list of Events with a label in the current namespace 02/25/23 10:59:35.482 + STEP: delete a list of events 02/25/23 10:59:35.507 + Feb 25 10:59:35.507: INFO: requesting DeleteCollection of events + STEP: check that the list of events matches the requested quantity 02/25/23 10:59:35.589 + [AfterEach] [sig-instrumentation] Events API test/e2e/framework/framework.go:187 - Dec 14 16:42:08.592: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "resourcequota-7588" for this suite. 12/14/22 16:42:08.598 + Feb 25 10:59:35.597: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "events-4930" for this suite. 02/25/23 10:59:35.628 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Secrets - should be consumable via the environment [NodeConformance] [Conformance] - test/e2e/common/node/secrets.go:94 -[BeforeEach] [sig-node] Secrets +[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] + should perform canary updates and phased rolling updates of template modifications [Conformance] + test/e2e/apps/statefulset.go:315 +[BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:42:08.611 -Dec 14 16:42:08.611: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename secrets 12/14/22 16:42:08.614 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:42:08.639 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:42:08.642 -[It] should be consumable via the environment [NodeConformance] [Conformance] - test/e2e/common/node/secrets.go:94 -STEP: creating secret secrets-1466/secret-test-2993544a-4c0d-4c6b-8c5a-09793a8ef986 12/14/22 16:42:08.646 -STEP: Creating a pod to test consume secrets 12/14/22 16:42:08.653 -Dec 14 16:42:08.662: INFO: Waiting up to 5m0s for pod "pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560" in namespace "secrets-1466" to be "Succeeded or Failed" -Dec 14 16:42:08.668: INFO: Pod "pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560": Phase="Pending", Reason="", readiness=false. Elapsed: 5.847867ms -Dec 14 16:42:10.678: INFO: Pod "pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015116141s -Dec 14 16:42:12.678: INFO: Pod "pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560": Phase="Pending", Reason="", readiness=false. Elapsed: 4.015411011s -Dec 14 16:42:14.676: INFO: Pod "pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.013394877s -STEP: Saw pod success 12/14/22 16:42:14.676 -Dec 14 16:42:14.676: INFO: Pod "pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560" satisfied condition "Succeeded or Failed" -Dec 14 16:42:14.681: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560 container env-test: -STEP: delete the pod 12/14/22 16:42:14.694 -Dec 14 16:42:14.716: INFO: Waiting for pod pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560 to disappear -Dec 14 16:42:14.721: INFO: Pod pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560 no longer exists -[AfterEach] [sig-node] Secrets +STEP: Creating a kubernetes client 02/25/23 10:59:35.698 +Feb 25 10:59:35.698: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename statefulset 02/25/23 10:59:35.705 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:59:35.758 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:59:35.77 +[BeforeEach] [sig-apps] StatefulSet + test/e2e/apps/statefulset.go:96 +[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:111 +STEP: Creating service test in namespace statefulset-7832 02/25/23 10:59:35.779 +[It] should perform canary updates and phased rolling updates of template modifications [Conformance] + test/e2e/apps/statefulset.go:315 +STEP: Creating a new StatefulSet 02/25/23 10:59:35.822 +Feb 25 10:59:35.859: INFO: Found 0 stateful pods, waiting for 3 +Feb 25 10:59:45.872: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true +Feb 25 10:59:45.873: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true +Feb 25 10:59:45.873: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true +STEP: Updating stateful set template: update image from registry.k8s.io/e2e-test-images/httpd:2.4.38-2 to registry.k8s.io/e2e-test-images/httpd:2.4.39-2 02/25/23 10:59:45.903 +Feb 25 10:59:45.942: INFO: Updating stateful set ss2 +STEP: Creating a new revision 02/25/23 10:59:45.943 +STEP: Not applying an update when the partition is greater than the number of replicas 02/25/23 10:59:55.991 +STEP: Performing a canary update 02/25/23 10:59:55.991 +Feb 25 10:59:56.022: INFO: Updating stateful set ss2 +Feb 25 10:59:56.039: INFO: Waiting for Pod statefulset-7832/ss2-2 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 +STEP: Restoring Pods to the correct revision when they are deleted 02/25/23 11:00:06.06 +Feb 25 11:00:06.191: INFO: Found 1 stateful pods, waiting for 3 +Feb 25 11:00:16.207: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true +Feb 25 11:00:16.208: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true +Feb 25 11:00:16.208: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true +STEP: Performing a phased rolling update 02/25/23 11:00:16.229 +Feb 25 11:00:16.258: INFO: Updating stateful set ss2 +Feb 25 11:00:16.276: INFO: Waiting for Pod statefulset-7832/ss2-1 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 +Feb 25 11:00:26.335: INFO: Updating stateful set ss2 +Feb 25 11:00:26.364: INFO: Waiting for StatefulSet statefulset-7832/ss2 to complete update +Feb 25 11:00:26.364: INFO: Waiting for Pod statefulset-7832/ss2-0 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 +[AfterEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:122 +Feb 25 11:00:36.405: INFO: Deleting all statefulset in ns statefulset-7832 +Feb 25 11:00:36.415: INFO: Scaling statefulset ss2 to 0 +Feb 25 11:00:46.474: INFO: Waiting for statefulset status.replicas updated to 0 +Feb 25 11:00:46.482: INFO: Deleting statefulset ss2 +[AfterEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:187 -Dec 14 16:42:14.722: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-1466" for this suite. 12/14/22 16:42:14.727 -{"msg":"PASSED [sig-node] Secrets should be consumable via the environment [NodeConformance] [Conformance]","completed":226,"skipped":4234,"failed":0} +Feb 25 11:00:46.530: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "statefulset-7832" for this suite. 02/25/23 11:00:46.55 +{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] should perform canary updates and phased rolling updates of template modifications [Conformance]","completed":222,"skipped":4074,"failed":0} ------------------------------ -• [SLOW TEST] [6.134 seconds] -[sig-node] Secrets -test/e2e/common/node/framework.go:23 - should be consumable via the environment [NodeConformance] [Conformance] - test/e2e/common/node/secrets.go:94 +• [SLOW TEST] [70.886 seconds] +[sig-apps] StatefulSet +test/e2e/apps/framework.go:23 + Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:101 + should perform canary updates and phased rolling updates of template modifications [Conformance] + test/e2e/apps/statefulset.go:315 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Secrets + [BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:42:08.611 - Dec 14 16:42:08.611: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename secrets 12/14/22 16:42:08.614 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:42:08.639 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:42:08.642 - [It] should be consumable via the environment [NodeConformance] [Conformance] - test/e2e/common/node/secrets.go:94 - STEP: creating secret secrets-1466/secret-test-2993544a-4c0d-4c6b-8c5a-09793a8ef986 12/14/22 16:42:08.646 - STEP: Creating a pod to test consume secrets 12/14/22 16:42:08.653 - Dec 14 16:42:08.662: INFO: Waiting up to 5m0s for pod "pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560" in namespace "secrets-1466" to be "Succeeded or Failed" - Dec 14 16:42:08.668: INFO: Pod "pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560": Phase="Pending", Reason="", readiness=false. Elapsed: 5.847867ms - Dec 14 16:42:10.678: INFO: Pod "pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015116141s - Dec 14 16:42:12.678: INFO: Pod "pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560": Phase="Pending", Reason="", readiness=false. Elapsed: 4.015411011s - Dec 14 16:42:14.676: INFO: Pod "pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.013394877s - STEP: Saw pod success 12/14/22 16:42:14.676 - Dec 14 16:42:14.676: INFO: Pod "pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560" satisfied condition "Succeeded or Failed" - Dec 14 16:42:14.681: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560 container env-test: - STEP: delete the pod 12/14/22 16:42:14.694 - Dec 14 16:42:14.716: INFO: Waiting for pod pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560 to disappear - Dec 14 16:42:14.721: INFO: Pod pod-configmaps-cacebc1d-e3cf-4569-80e7-0feaed11d560 no longer exists - [AfterEach] [sig-node] Secrets + STEP: Creating a kubernetes client 02/25/23 10:59:35.698 + Feb 25 10:59:35.698: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename statefulset 02/25/23 10:59:35.705 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 10:59:35.758 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 10:59:35.77 + [BeforeEach] [sig-apps] StatefulSet + test/e2e/apps/statefulset.go:96 + [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:111 + STEP: Creating service test in namespace statefulset-7832 02/25/23 10:59:35.779 + [It] should perform canary updates and phased rolling updates of template modifications [Conformance] + test/e2e/apps/statefulset.go:315 + STEP: Creating a new StatefulSet 02/25/23 10:59:35.822 + Feb 25 10:59:35.859: INFO: Found 0 stateful pods, waiting for 3 + Feb 25 10:59:45.872: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true + Feb 25 10:59:45.873: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true + Feb 25 10:59:45.873: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true + STEP: Updating stateful set template: update image from registry.k8s.io/e2e-test-images/httpd:2.4.38-2 to registry.k8s.io/e2e-test-images/httpd:2.4.39-2 02/25/23 10:59:45.903 + Feb 25 10:59:45.942: INFO: Updating stateful set ss2 + STEP: Creating a new revision 02/25/23 10:59:45.943 + STEP: Not applying an update when the partition is greater than the number of replicas 02/25/23 10:59:55.991 + STEP: Performing a canary update 02/25/23 10:59:55.991 + Feb 25 10:59:56.022: INFO: Updating stateful set ss2 + Feb 25 10:59:56.039: INFO: Waiting for Pod statefulset-7832/ss2-2 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 + STEP: Restoring Pods to the correct revision when they are deleted 02/25/23 11:00:06.06 + Feb 25 11:00:06.191: INFO: Found 1 stateful pods, waiting for 3 + Feb 25 11:00:16.207: INFO: Waiting for pod ss2-0 to enter Running - Ready=true, currently Running - Ready=true + Feb 25 11:00:16.208: INFO: Waiting for pod ss2-1 to enter Running - Ready=true, currently Running - Ready=true + Feb 25 11:00:16.208: INFO: Waiting for pod ss2-2 to enter Running - Ready=true, currently Running - Ready=true + STEP: Performing a phased rolling update 02/25/23 11:00:16.229 + Feb 25 11:00:16.258: INFO: Updating stateful set ss2 + Feb 25 11:00:16.276: INFO: Waiting for Pod statefulset-7832/ss2-1 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 + Feb 25 11:00:26.335: INFO: Updating stateful set ss2 + Feb 25 11:00:26.364: INFO: Waiting for StatefulSet statefulset-7832/ss2 to complete update + Feb 25 11:00:26.364: INFO: Waiting for Pod statefulset-7832/ss2-0 to have revision ss2-5d8c6ff87d update revision ss2-6557876d87 + [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:122 + Feb 25 11:00:36.405: INFO: Deleting all statefulset in ns statefulset-7832 + Feb 25 11:00:36.415: INFO: Scaling statefulset ss2 to 0 + Feb 25 11:00:46.474: INFO: Waiting for statefulset status.replicas updated to 0 + Feb 25 11:00:46.482: INFO: Deleting statefulset ss2 + [AfterEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:187 - Dec 14 16:42:14.722: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "secrets-1466" for this suite. 12/14/22 16:42:14.727 + Feb 25 11:00:46.530: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "statefulset-7832" for this suite. 02/25/23 11:00:46.55 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Downward API volume - should provide podname only [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:52 -[BeforeEach] [sig-storage] Downward API volume +[sig-node] Containers + should be able to override the image's default command and arguments [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:86 +[BeforeEach] [sig-node] Containers test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:42:14.755 -Dec 14 16:42:14.756: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 16:42:14.759 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:42:14.782 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:42:14.785 -[BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 -[It] should provide podname only [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:52 -STEP: Creating a pod to test downward API volume plugin 12/14/22 16:42:14.789 -Dec 14 16:42:14.800: INFO: Waiting up to 5m0s for pod "downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d" in namespace "downward-api-9325" to be "Succeeded or Failed" -Dec 14 16:42:14.806: INFO: Pod "downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d": Phase="Pending", Reason="", readiness=false. Elapsed: 5.739135ms -Dec 14 16:42:16.814: INFO: Pod "downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014513153s -Dec 14 16:42:18.812: INFO: Pod "downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d": Phase="Pending", Reason="", readiness=false. Elapsed: 4.01248073s -Dec 14 16:42:20.812: INFO: Pod "downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.011979954s -STEP: Saw pod success 12/14/22 16:42:20.812 -Dec 14 16:42:20.812: INFO: Pod "downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d" satisfied condition "Succeeded or Failed" -Dec 14 16:42:20.819: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d container client-container: -STEP: delete the pod 12/14/22 16:42:20.828 -Dec 14 16:42:20.847: INFO: Waiting for pod downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d to disappear -Dec 14 16:42:20.852: INFO: Pod downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d no longer exists -[AfterEach] [sig-storage] Downward API volume +STEP: Creating a kubernetes client 02/25/23 11:00:46.604 +Feb 25 11:00:46.604: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename containers 02/25/23 11:00:46.607 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:00:46.637 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:00:46.643 +[It] should be able to override the image's default command and arguments [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:86 +STEP: Creating a pod to test override all 02/25/23 11:00:46.653 +Feb 25 11:00:46.668: INFO: Waiting up to 5m0s for pod "client-containers-2053c64a-3748-4544-ac48-b893218094a0" in namespace "containers-1664" to be "Succeeded or Failed" +Feb 25 11:00:46.678: INFO: Pod "client-containers-2053c64a-3748-4544-ac48-b893218094a0": Phase="Pending", Reason="", readiness=false. Elapsed: 9.559916ms +Feb 25 11:00:48.690: INFO: Pod "client-containers-2053c64a-3748-4544-ac48-b893218094a0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022063831s +Feb 25 11:00:50.692: INFO: Pod "client-containers-2053c64a-3748-4544-ac48-b893218094a0": Phase="Pending", Reason="", readiness=false. Elapsed: 4.023606851s +Feb 25 11:00:52.685: INFO: Pod "client-containers-2053c64a-3748-4544-ac48-b893218094a0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016725104s +STEP: Saw pod success 02/25/23 11:00:52.685 +Feb 25 11:00:52.685: INFO: Pod "client-containers-2053c64a-3748-4544-ac48-b893218094a0" satisfied condition "Succeeded or Failed" +Feb 25 11:00:52.711: INFO: Trying to get logs from node diek3dua9gay-3 pod client-containers-2053c64a-3748-4544-ac48-b893218094a0 container agnhost-container: +STEP: delete the pod 02/25/23 11:00:52.735 +Feb 25 11:00:52.782: INFO: Waiting for pod client-containers-2053c64a-3748-4544-ac48-b893218094a0 to disappear +Feb 25 11:00:52.803: INFO: Pod client-containers-2053c64a-3748-4544-ac48-b893218094a0 no longer exists +[AfterEach] [sig-node] Containers test/e2e/framework/framework.go:187 -Dec 14 16:42:20.852: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-9325" for this suite. 12/14/22 16:42:20.862 -{"msg":"PASSED [sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance]","completed":227,"skipped":4270,"failed":0} +Feb 25 11:00:52.804: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "containers-1664" for this suite. 02/25/23 11:00:52.833 +{"msg":"PASSED [sig-node] Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance]","completed":223,"skipped":4138,"failed":0} ------------------------------ -• [SLOW TEST] [6.114 seconds] -[sig-storage] Downward API volume -test/e2e/common/storage/framework.go:23 - should provide podname only [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:52 +• [SLOW TEST] [6.244 seconds] +[sig-node] Containers +test/e2e/common/node/framework.go:23 + should be able to override the image's default command and arguments [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:86 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Downward API volume + [BeforeEach] [sig-node] Containers test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:42:14.755 - Dec 14 16:42:14.756: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 16:42:14.759 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:42:14.782 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:42:14.785 - [BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 - [It] should provide podname only [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:52 - STEP: Creating a pod to test downward API volume plugin 12/14/22 16:42:14.789 - Dec 14 16:42:14.800: INFO: Waiting up to 5m0s for pod "downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d" in namespace "downward-api-9325" to be "Succeeded or Failed" - Dec 14 16:42:14.806: INFO: Pod "downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d": Phase="Pending", Reason="", readiness=false. Elapsed: 5.739135ms - Dec 14 16:42:16.814: INFO: Pod "downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014513153s - Dec 14 16:42:18.812: INFO: Pod "downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d": Phase="Pending", Reason="", readiness=false. Elapsed: 4.01248073s - Dec 14 16:42:20.812: INFO: Pod "downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.011979954s - STEP: Saw pod success 12/14/22 16:42:20.812 - Dec 14 16:42:20.812: INFO: Pod "downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d" satisfied condition "Succeeded or Failed" - Dec 14 16:42:20.819: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d container client-container: - STEP: delete the pod 12/14/22 16:42:20.828 - Dec 14 16:42:20.847: INFO: Waiting for pod downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d to disappear - Dec 14 16:42:20.852: INFO: Pod downwardapi-volume-c766007e-ddc8-4e3d-be69-2e8636c02b0d no longer exists - [AfterEach] [sig-storage] Downward API volume + STEP: Creating a kubernetes client 02/25/23 11:00:46.604 + Feb 25 11:00:46.604: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename containers 02/25/23 11:00:46.607 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:00:46.637 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:00:46.643 + [It] should be able to override the image's default command and arguments [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:86 + STEP: Creating a pod to test override all 02/25/23 11:00:46.653 + Feb 25 11:00:46.668: INFO: Waiting up to 5m0s for pod "client-containers-2053c64a-3748-4544-ac48-b893218094a0" in namespace "containers-1664" to be "Succeeded or Failed" + Feb 25 11:00:46.678: INFO: Pod "client-containers-2053c64a-3748-4544-ac48-b893218094a0": Phase="Pending", Reason="", readiness=false. Elapsed: 9.559916ms + Feb 25 11:00:48.690: INFO: Pod "client-containers-2053c64a-3748-4544-ac48-b893218094a0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022063831s + Feb 25 11:00:50.692: INFO: Pod "client-containers-2053c64a-3748-4544-ac48-b893218094a0": Phase="Pending", Reason="", readiness=false. Elapsed: 4.023606851s + Feb 25 11:00:52.685: INFO: Pod "client-containers-2053c64a-3748-4544-ac48-b893218094a0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016725104s + STEP: Saw pod success 02/25/23 11:00:52.685 + Feb 25 11:00:52.685: INFO: Pod "client-containers-2053c64a-3748-4544-ac48-b893218094a0" satisfied condition "Succeeded or Failed" + Feb 25 11:00:52.711: INFO: Trying to get logs from node diek3dua9gay-3 pod client-containers-2053c64a-3748-4544-ac48-b893218094a0 container agnhost-container: + STEP: delete the pod 02/25/23 11:00:52.735 + Feb 25 11:00:52.782: INFO: Waiting for pod client-containers-2053c64a-3748-4544-ac48-b893218094a0 to disappear + Feb 25 11:00:52.803: INFO: Pod client-containers-2053c64a-3748-4544-ac48-b893218094a0 no longer exists + [AfterEach] [sig-node] Containers test/e2e/framework/framework.go:187 - Dec 14 16:42:20.852: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-9325" for this suite. 12/14/22 16:42:20.862 + Feb 25 11:00:52.804: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "containers-1664" for this suite. 02/25/23 11:00:52.833 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSS ------------------------------ -[sig-instrumentation] Events API - should ensure that an event can be fetched, patched, deleted, and listed [Conformance] - test/e2e/instrumentation/events.go:98 -[BeforeEach] [sig-instrumentation] Events API +[sig-network] HostPort + validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance] + test/e2e/network/hostport.go:63 +[BeforeEach] [sig-network] HostPort test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:42:20.88 -Dec 14 16:42:20.880: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename events 12/14/22 16:42:20.882 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:42:20.913 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:42:20.918 -[BeforeEach] [sig-instrumentation] Events API - test/e2e/instrumentation/events.go:84 -[It] should ensure that an event can be fetched, patched, deleted, and listed [Conformance] - test/e2e/instrumentation/events.go:98 -STEP: creating a test event 12/14/22 16:42:20.924 -STEP: listing events in all namespaces 12/14/22 16:42:20.935 -STEP: listing events in test namespace 12/14/22 16:42:20.942 -STEP: listing events with field selection filtering on source 12/14/22 16:42:20.947 -STEP: listing events with field selection filtering on reportingController 12/14/22 16:42:20.952 -STEP: getting the test event 12/14/22 16:42:20.956 -STEP: patching the test event 12/14/22 16:42:20.96 -STEP: getting the test event 12/14/22 16:42:20.978 -STEP: updating the test event 12/14/22 16:42:20.982 -STEP: getting the test event 12/14/22 16:42:20.992 -STEP: deleting the test event 12/14/22 16:42:20.997 -STEP: listing events in all namespaces 12/14/22 16:42:21.008 -STEP: listing events in test namespace 12/14/22 16:42:21.014 -[AfterEach] [sig-instrumentation] Events API +STEP: Creating a kubernetes client 02/25/23 11:00:52.853 +Feb 25 11:00:52.853: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename hostport 02/25/23 11:00:52.858 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:00:53.051 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:00:53.06 +[BeforeEach] [sig-network] HostPort + test/e2e/network/hostport.go:49 +[It] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance] + test/e2e/network/hostport.go:63 +STEP: Trying to create a pod(pod1) with hostport 54323 and hostIP 127.0.0.1 and expect scheduled 02/25/23 11:00:53.161 +Feb 25 11:00:53.294: INFO: Waiting up to 5m0s for pod "pod1" in namespace "hostport-6934" to be "running and ready" +Feb 25 11:00:53.360: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 65.643468ms +Feb 25 11:00:53.360: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:00:55.370: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.075651998s +Feb 25 11:00:55.370: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:00:57.372: INFO: Pod "pod1": Phase="Running", Reason="", readiness=true. Elapsed: 4.07768917s +Feb 25 11:00:57.372: INFO: The phase of Pod pod1 is Running (Ready = true) +Feb 25 11:00:57.372: INFO: Pod "pod1" satisfied condition "running and ready" +STEP: Trying to create another pod(pod2) with hostport 54323 but hostIP 192.168.121.117 on the node which pod1 resides and expect scheduled 02/25/23 11:00:57.372 +Feb 25 11:00:57.386: INFO: Waiting up to 5m0s for pod "pod2" in namespace "hostport-6934" to be "running and ready" +Feb 25 11:00:57.398: INFO: Pod "pod2": Phase="Pending", Reason="", readiness=false. Elapsed: 11.161041ms +Feb 25 11:00:57.398: INFO: The phase of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:00:59.411: INFO: Pod "pod2": Phase="Running", Reason="", readiness=false. Elapsed: 2.023946756s +Feb 25 11:00:59.411: INFO: The phase of Pod pod2 is Running (Ready = false) +Feb 25 11:01:01.407: INFO: Pod "pod2": Phase="Running", Reason="", readiness=true. Elapsed: 4.020563656s +Feb 25 11:01:01.407: INFO: The phase of Pod pod2 is Running (Ready = true) +Feb 25 11:01:01.407: INFO: Pod "pod2" satisfied condition "running and ready" +STEP: Trying to create a third pod(pod3) with hostport 54323, hostIP 192.168.121.117 but use UDP protocol on the node which pod2 resides 02/25/23 11:01:01.408 +Feb 25 11:01:01.422: INFO: Waiting up to 5m0s for pod "pod3" in namespace "hostport-6934" to be "running and ready" +Feb 25 11:01:01.434: INFO: Pod "pod3": Phase="Pending", Reason="", readiness=false. Elapsed: 12.558368ms +Feb 25 11:01:01.434: INFO: The phase of Pod pod3 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:01:03.452: INFO: Pod "pod3": Phase="Running", Reason="", readiness=true. Elapsed: 2.030692709s +Feb 25 11:01:03.452: INFO: The phase of Pod pod3 is Running (Ready = true) +Feb 25 11:01:03.452: INFO: Pod "pod3" satisfied condition "running and ready" +Feb 25 11:01:03.472: INFO: Waiting up to 5m0s for pod "e2e-host-exec" in namespace "hostport-6934" to be "running and ready" +Feb 25 11:01:03.480: INFO: Pod "e2e-host-exec": Phase="Pending", Reason="", readiness=false. Elapsed: 8.114067ms +Feb 25 11:01:03.480: INFO: The phase of Pod e2e-host-exec is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:01:05.488: INFO: Pod "e2e-host-exec": Phase="Running", Reason="", readiness=true. Elapsed: 2.016359375s +Feb 25 11:01:05.489: INFO: The phase of Pod e2e-host-exec is Running (Ready = true) +Feb 25 11:01:05.489: INFO: Pod "e2e-host-exec" satisfied condition "running and ready" +STEP: checking connectivity from pod e2e-host-exec to serverIP: 127.0.0.1, port: 54323 02/25/23 11:01:05.495 +Feb 25 11:01:05.496: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g --connect-timeout 5 --interface 192.168.121.117 http://127.0.0.1:54323/hostname] Namespace:hostport-6934 PodName:e2e-host-exec ContainerName:e2e-host-exec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 11:01:05.496: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 11:01:05.498: INFO: ExecWithOptions: Clientset creation +Feb 25 11:01:05.498: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/hostport-6934/pods/e2e-host-exec/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+--connect-timeout+5+--interface+192.168.121.117+http%3A%2F%2F127.0.0.1%3A54323%2Fhostname&container=e2e-host-exec&container=e2e-host-exec&stderr=true&stdout=true) +STEP: checking connectivity from pod e2e-host-exec to serverIP: 192.168.121.117, port: 54323 02/25/23 11:01:05.707 +Feb 25 11:01:05.707: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g --connect-timeout 5 http://192.168.121.117:54323/hostname] Namespace:hostport-6934 PodName:e2e-host-exec ContainerName:e2e-host-exec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 11:01:05.707: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 11:01:05.711: INFO: ExecWithOptions: Clientset creation +Feb 25 11:01:05.711: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/hostport-6934/pods/e2e-host-exec/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+--connect-timeout+5+http%3A%2F%2F192.168.121.117%3A54323%2Fhostname&container=e2e-host-exec&container=e2e-host-exec&stderr=true&stdout=true) +STEP: checking connectivity from pod e2e-host-exec to serverIP: 192.168.121.117, port: 54323 UDP 02/25/23 11:01:05.88 +Feb 25 11:01:05.881: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostname | nc -u -w 5 192.168.121.117 54323] Namespace:hostport-6934 PodName:e2e-host-exec ContainerName:e2e-host-exec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 11:01:05.881: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 11:01:05.882: INFO: ExecWithOptions: Clientset creation +Feb 25 11:01:05.882: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/hostport-6934/pods/e2e-host-exec/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostname+%7C+nc+-u+-w+5+192.168.121.117+54323&container=e2e-host-exec&container=e2e-host-exec&stderr=true&stdout=true) +[AfterEach] [sig-network] HostPort test/e2e/framework/framework.go:187 -Dec 14 16:42:21.017: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "events-3594" for this suite. 12/14/22 16:42:21.023 -{"msg":"PASSED [sig-instrumentation] Events API should ensure that an event can be fetched, patched, deleted, and listed [Conformance]","completed":228,"skipped":4300,"failed":0} +Feb 25 11:01:11.029: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "hostport-6934" for this suite. 02/25/23 11:01:11.038 +{"msg":"PASSED [sig-network] HostPort validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance]","completed":224,"skipped":4145,"failed":0} ------------------------------ -• [0.152 seconds] -[sig-instrumentation] Events API -test/e2e/instrumentation/common/framework.go:23 - should ensure that an event can be fetched, patched, deleted, and listed [Conformance] - test/e2e/instrumentation/events.go:98 +• [SLOW TEST] [18.205 seconds] +[sig-network] HostPort +test/e2e/network/common/framework.go:23 + validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance] + test/e2e/network/hostport.go:63 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-instrumentation] Events API - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:42:20.88 - Dec 14 16:42:20.880: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename events 12/14/22 16:42:20.882 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:42:20.913 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:42:20.918 - [BeforeEach] [sig-instrumentation] Events API - test/e2e/instrumentation/events.go:84 - [It] should ensure that an event can be fetched, patched, deleted, and listed [Conformance] - test/e2e/instrumentation/events.go:98 - STEP: creating a test event 12/14/22 16:42:20.924 - STEP: listing events in all namespaces 12/14/22 16:42:20.935 - STEP: listing events in test namespace 12/14/22 16:42:20.942 - STEP: listing events with field selection filtering on source 12/14/22 16:42:20.947 - STEP: listing events with field selection filtering on reportingController 12/14/22 16:42:20.952 - STEP: getting the test event 12/14/22 16:42:20.956 - STEP: patching the test event 12/14/22 16:42:20.96 - STEP: getting the test event 12/14/22 16:42:20.978 - STEP: updating the test event 12/14/22 16:42:20.982 - STEP: getting the test event 12/14/22 16:42:20.992 - STEP: deleting the test event 12/14/22 16:42:20.997 - STEP: listing events in all namespaces 12/14/22 16:42:21.008 - STEP: listing events in test namespace 12/14/22 16:42:21.014 - [AfterEach] [sig-instrumentation] Events API + [BeforeEach] [sig-network] HostPort + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 11:00:52.853 + Feb 25 11:00:52.853: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename hostport 02/25/23 11:00:52.858 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:00:53.051 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:00:53.06 + [BeforeEach] [sig-network] HostPort + test/e2e/network/hostport.go:49 + [It] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance] + test/e2e/network/hostport.go:63 + STEP: Trying to create a pod(pod1) with hostport 54323 and hostIP 127.0.0.1 and expect scheduled 02/25/23 11:00:53.161 + Feb 25 11:00:53.294: INFO: Waiting up to 5m0s for pod "pod1" in namespace "hostport-6934" to be "running and ready" + Feb 25 11:00:53.360: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 65.643468ms + Feb 25 11:00:53.360: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:00:55.370: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.075651998s + Feb 25 11:00:55.370: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:00:57.372: INFO: Pod "pod1": Phase="Running", Reason="", readiness=true. Elapsed: 4.07768917s + Feb 25 11:00:57.372: INFO: The phase of Pod pod1 is Running (Ready = true) + Feb 25 11:00:57.372: INFO: Pod "pod1" satisfied condition "running and ready" + STEP: Trying to create another pod(pod2) with hostport 54323 but hostIP 192.168.121.117 on the node which pod1 resides and expect scheduled 02/25/23 11:00:57.372 + Feb 25 11:00:57.386: INFO: Waiting up to 5m0s for pod "pod2" in namespace "hostport-6934" to be "running and ready" + Feb 25 11:00:57.398: INFO: Pod "pod2": Phase="Pending", Reason="", readiness=false. Elapsed: 11.161041ms + Feb 25 11:00:57.398: INFO: The phase of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:00:59.411: INFO: Pod "pod2": Phase="Running", Reason="", readiness=false. Elapsed: 2.023946756s + Feb 25 11:00:59.411: INFO: The phase of Pod pod2 is Running (Ready = false) + Feb 25 11:01:01.407: INFO: Pod "pod2": Phase="Running", Reason="", readiness=true. Elapsed: 4.020563656s + Feb 25 11:01:01.407: INFO: The phase of Pod pod2 is Running (Ready = true) + Feb 25 11:01:01.407: INFO: Pod "pod2" satisfied condition "running and ready" + STEP: Trying to create a third pod(pod3) with hostport 54323, hostIP 192.168.121.117 but use UDP protocol on the node which pod2 resides 02/25/23 11:01:01.408 + Feb 25 11:01:01.422: INFO: Waiting up to 5m0s for pod "pod3" in namespace "hostport-6934" to be "running and ready" + Feb 25 11:01:01.434: INFO: Pod "pod3": Phase="Pending", Reason="", readiness=false. Elapsed: 12.558368ms + Feb 25 11:01:01.434: INFO: The phase of Pod pod3 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:01:03.452: INFO: Pod "pod3": Phase="Running", Reason="", readiness=true. Elapsed: 2.030692709s + Feb 25 11:01:03.452: INFO: The phase of Pod pod3 is Running (Ready = true) + Feb 25 11:01:03.452: INFO: Pod "pod3" satisfied condition "running and ready" + Feb 25 11:01:03.472: INFO: Waiting up to 5m0s for pod "e2e-host-exec" in namespace "hostport-6934" to be "running and ready" + Feb 25 11:01:03.480: INFO: Pod "e2e-host-exec": Phase="Pending", Reason="", readiness=false. Elapsed: 8.114067ms + Feb 25 11:01:03.480: INFO: The phase of Pod e2e-host-exec is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:01:05.488: INFO: Pod "e2e-host-exec": Phase="Running", Reason="", readiness=true. Elapsed: 2.016359375s + Feb 25 11:01:05.489: INFO: The phase of Pod e2e-host-exec is Running (Ready = true) + Feb 25 11:01:05.489: INFO: Pod "e2e-host-exec" satisfied condition "running and ready" + STEP: checking connectivity from pod e2e-host-exec to serverIP: 127.0.0.1, port: 54323 02/25/23 11:01:05.495 + Feb 25 11:01:05.496: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g --connect-timeout 5 --interface 192.168.121.117 http://127.0.0.1:54323/hostname] Namespace:hostport-6934 PodName:e2e-host-exec ContainerName:e2e-host-exec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 11:01:05.496: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 11:01:05.498: INFO: ExecWithOptions: Clientset creation + Feb 25 11:01:05.498: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/hostport-6934/pods/e2e-host-exec/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+--connect-timeout+5+--interface+192.168.121.117+http%3A%2F%2F127.0.0.1%3A54323%2Fhostname&container=e2e-host-exec&container=e2e-host-exec&stderr=true&stdout=true) + STEP: checking connectivity from pod e2e-host-exec to serverIP: 192.168.121.117, port: 54323 02/25/23 11:01:05.707 + Feb 25 11:01:05.707: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g --connect-timeout 5 http://192.168.121.117:54323/hostname] Namespace:hostport-6934 PodName:e2e-host-exec ContainerName:e2e-host-exec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 11:01:05.707: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 11:01:05.711: INFO: ExecWithOptions: Clientset creation + Feb 25 11:01:05.711: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/hostport-6934/pods/e2e-host-exec/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+--connect-timeout+5+http%3A%2F%2F192.168.121.117%3A54323%2Fhostname&container=e2e-host-exec&container=e2e-host-exec&stderr=true&stdout=true) + STEP: checking connectivity from pod e2e-host-exec to serverIP: 192.168.121.117, port: 54323 UDP 02/25/23 11:01:05.88 + Feb 25 11:01:05.881: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostname | nc -u -w 5 192.168.121.117 54323] Namespace:hostport-6934 PodName:e2e-host-exec ContainerName:e2e-host-exec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 11:01:05.881: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 11:01:05.882: INFO: ExecWithOptions: Clientset creation + Feb 25 11:01:05.882: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/hostport-6934/pods/e2e-host-exec/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostname+%7C+nc+-u+-w+5+192.168.121.117+54323&container=e2e-host-exec&container=e2e-host-exec&stderr=true&stdout=true) + [AfterEach] [sig-network] HostPort test/e2e/framework/framework.go:187 - Dec 14 16:42:21.017: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "events-3594" for this suite. 12/14/22 16:42:21.023 + Feb 25 11:01:11.029: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "hostport-6934" for this suite. 02/25/23 11:01:11.038 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSS +S ------------------------------ -[sig-api-machinery] Discovery - should validate PreferredVersion for each APIGroup [Conformance] - test/e2e/apimachinery/discovery.go:122 -[BeforeEach] [sig-api-machinery] Discovery +[sig-storage] Projected downwardAPI + should update labels on modification [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:129 +[BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:42:21.034 -Dec 14 16:42:21.034: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename discovery 12/14/22 16:42:21.036 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:42:21.058 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:42:21.064 -[BeforeEach] [sig-api-machinery] Discovery - test/e2e/apimachinery/discovery.go:43 -STEP: Setting up server cert 12/14/22 16:42:21.067 -[It] should validate PreferredVersion for each APIGroup [Conformance] - test/e2e/apimachinery/discovery.go:122 -Dec 14 16:42:21.674: INFO: Checking APIGroup: apiregistration.k8s.io -Dec 14 16:42:21.676: INFO: PreferredVersion.GroupVersion: apiregistration.k8s.io/v1 -Dec 14 16:42:21.676: INFO: Versions found [{apiregistration.k8s.io/v1 v1}] -Dec 14 16:42:21.676: INFO: apiregistration.k8s.io/v1 matches apiregistration.k8s.io/v1 -Dec 14 16:42:21.676: INFO: Checking APIGroup: apps -Dec 14 16:42:21.678: INFO: PreferredVersion.GroupVersion: apps/v1 -Dec 14 16:42:21.678: INFO: Versions found [{apps/v1 v1}] -Dec 14 16:42:21.678: INFO: apps/v1 matches apps/v1 -Dec 14 16:42:21.678: INFO: Checking APIGroup: events.k8s.io -Dec 14 16:42:21.679: INFO: PreferredVersion.GroupVersion: events.k8s.io/v1 -Dec 14 16:42:21.679: INFO: Versions found [{events.k8s.io/v1 v1}] -Dec 14 16:42:21.679: INFO: events.k8s.io/v1 matches events.k8s.io/v1 -Dec 14 16:42:21.679: INFO: Checking APIGroup: authentication.k8s.io -Dec 14 16:42:21.681: INFO: PreferredVersion.GroupVersion: authentication.k8s.io/v1 -Dec 14 16:42:21.681: INFO: Versions found [{authentication.k8s.io/v1 v1}] -Dec 14 16:42:21.681: INFO: authentication.k8s.io/v1 matches authentication.k8s.io/v1 -Dec 14 16:42:21.681: INFO: Checking APIGroup: authorization.k8s.io -Dec 14 16:42:21.682: INFO: PreferredVersion.GroupVersion: authorization.k8s.io/v1 -Dec 14 16:42:21.682: INFO: Versions found [{authorization.k8s.io/v1 v1}] -Dec 14 16:42:21.682: INFO: authorization.k8s.io/v1 matches authorization.k8s.io/v1 -Dec 14 16:42:21.682: INFO: Checking APIGroup: autoscaling -Dec 14 16:42:21.684: INFO: PreferredVersion.GroupVersion: autoscaling/v2 -Dec 14 16:42:21.684: INFO: Versions found [{autoscaling/v2 v2} {autoscaling/v1 v1} {autoscaling/v2beta2 v2beta2}] -Dec 14 16:42:21.684: INFO: autoscaling/v2 matches autoscaling/v2 -Dec 14 16:42:21.684: INFO: Checking APIGroup: batch -Dec 14 16:42:21.685: INFO: PreferredVersion.GroupVersion: batch/v1 -Dec 14 16:42:21.685: INFO: Versions found [{batch/v1 v1}] -Dec 14 16:42:21.685: INFO: batch/v1 matches batch/v1 -Dec 14 16:42:21.685: INFO: Checking APIGroup: certificates.k8s.io -Dec 14 16:42:21.687: INFO: PreferredVersion.GroupVersion: certificates.k8s.io/v1 -Dec 14 16:42:21.687: INFO: Versions found [{certificates.k8s.io/v1 v1}] -Dec 14 16:42:21.687: INFO: certificates.k8s.io/v1 matches certificates.k8s.io/v1 -Dec 14 16:42:21.687: INFO: Checking APIGroup: networking.k8s.io -Dec 14 16:42:21.688: INFO: PreferredVersion.GroupVersion: networking.k8s.io/v1 -Dec 14 16:42:21.688: INFO: Versions found [{networking.k8s.io/v1 v1}] -Dec 14 16:42:21.688: INFO: networking.k8s.io/v1 matches networking.k8s.io/v1 -Dec 14 16:42:21.688: INFO: Checking APIGroup: policy -Dec 14 16:42:21.690: INFO: PreferredVersion.GroupVersion: policy/v1 -Dec 14 16:42:21.690: INFO: Versions found [{policy/v1 v1}] -Dec 14 16:42:21.690: INFO: policy/v1 matches policy/v1 -Dec 14 16:42:21.690: INFO: Checking APIGroup: rbac.authorization.k8s.io -Dec 14 16:42:21.691: INFO: PreferredVersion.GroupVersion: rbac.authorization.k8s.io/v1 -Dec 14 16:42:21.691: INFO: Versions found [{rbac.authorization.k8s.io/v1 v1}] -Dec 14 16:42:21.691: INFO: rbac.authorization.k8s.io/v1 matches rbac.authorization.k8s.io/v1 -Dec 14 16:42:21.691: INFO: Checking APIGroup: storage.k8s.io -Dec 14 16:42:21.692: INFO: PreferredVersion.GroupVersion: storage.k8s.io/v1 -Dec 14 16:42:21.692: INFO: Versions found [{storage.k8s.io/v1 v1} {storage.k8s.io/v1beta1 v1beta1}] -Dec 14 16:42:21.692: INFO: storage.k8s.io/v1 matches storage.k8s.io/v1 -Dec 14 16:42:21.692: INFO: Checking APIGroup: admissionregistration.k8s.io -Dec 14 16:42:21.694: INFO: PreferredVersion.GroupVersion: admissionregistration.k8s.io/v1 -Dec 14 16:42:21.694: INFO: Versions found [{admissionregistration.k8s.io/v1 v1}] -Dec 14 16:42:21.694: INFO: admissionregistration.k8s.io/v1 matches admissionregistration.k8s.io/v1 -Dec 14 16:42:21.694: INFO: Checking APIGroup: apiextensions.k8s.io -Dec 14 16:42:21.695: INFO: PreferredVersion.GroupVersion: apiextensions.k8s.io/v1 -Dec 14 16:42:21.695: INFO: Versions found [{apiextensions.k8s.io/v1 v1}] -Dec 14 16:42:21.695: INFO: apiextensions.k8s.io/v1 matches apiextensions.k8s.io/v1 -Dec 14 16:42:21.695: INFO: Checking APIGroup: scheduling.k8s.io -Dec 14 16:42:21.697: INFO: PreferredVersion.GroupVersion: scheduling.k8s.io/v1 -Dec 14 16:42:21.697: INFO: Versions found [{scheduling.k8s.io/v1 v1}] -Dec 14 16:42:21.697: INFO: scheduling.k8s.io/v1 matches scheduling.k8s.io/v1 -Dec 14 16:42:21.697: INFO: Checking APIGroup: coordination.k8s.io -Dec 14 16:42:21.698: INFO: PreferredVersion.GroupVersion: coordination.k8s.io/v1 -Dec 14 16:42:21.698: INFO: Versions found [{coordination.k8s.io/v1 v1}] -Dec 14 16:42:21.698: INFO: coordination.k8s.io/v1 matches coordination.k8s.io/v1 -Dec 14 16:42:21.698: INFO: Checking APIGroup: node.k8s.io -Dec 14 16:42:21.699: INFO: PreferredVersion.GroupVersion: node.k8s.io/v1 -Dec 14 16:42:21.699: INFO: Versions found [{node.k8s.io/v1 v1}] -Dec 14 16:42:21.699: INFO: node.k8s.io/v1 matches node.k8s.io/v1 -Dec 14 16:42:21.699: INFO: Checking APIGroup: discovery.k8s.io -Dec 14 16:42:21.700: INFO: PreferredVersion.GroupVersion: discovery.k8s.io/v1 -Dec 14 16:42:21.700: INFO: Versions found [{discovery.k8s.io/v1 v1}] -Dec 14 16:42:21.700: INFO: discovery.k8s.io/v1 matches discovery.k8s.io/v1 -Dec 14 16:42:21.700: INFO: Checking APIGroup: flowcontrol.apiserver.k8s.io -Dec 14 16:42:21.702: INFO: PreferredVersion.GroupVersion: flowcontrol.apiserver.k8s.io/v1beta2 -Dec 14 16:42:21.702: INFO: Versions found [{flowcontrol.apiserver.k8s.io/v1beta2 v1beta2} {flowcontrol.apiserver.k8s.io/v1beta1 v1beta1}] -Dec 14 16:42:21.702: INFO: flowcontrol.apiserver.k8s.io/v1beta2 matches flowcontrol.apiserver.k8s.io/v1beta2 -[AfterEach] [sig-api-machinery] Discovery +STEP: Creating a kubernetes client 02/25/23 11:01:11.064 +Feb 25 11:01:11.064: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 11:01:11.07 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:01:11.103 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:01:11.109 +[BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 +[It] should update labels on modification [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:129 +STEP: Creating the pod 02/25/23 11:01:11.115 +Feb 25 11:01:11.133: INFO: Waiting up to 5m0s for pod "labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9" in namespace "projected-6214" to be "running and ready" +Feb 25 11:01:11.143: INFO: Pod "labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9": Phase="Pending", Reason="", readiness=false. Elapsed: 9.892354ms +Feb 25 11:01:11.143: INFO: The phase of Pod labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:01:13.164: INFO: Pod "labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031280921s +Feb 25 11:01:13.164: INFO: The phase of Pod labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:01:15.151: INFO: Pod "labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9": Phase="Running", Reason="", readiness=true. Elapsed: 4.017752184s +Feb 25 11:01:15.151: INFO: The phase of Pod labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9 is Running (Ready = true) +Feb 25 11:01:15.151: INFO: Pod "labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9" satisfied condition "running and ready" +Feb 25 11:01:15.699: INFO: Successfully updated pod "labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9" +[AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 -Dec 14 16:42:21.702: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "discovery-9662" for this suite. 12/14/22 16:42:21.709 -{"msg":"PASSED [sig-api-machinery] Discovery should validate PreferredVersion for each APIGroup [Conformance]","completed":229,"skipped":4314,"failed":0} +Feb 25 11:01:17.947: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-6214" for this suite. 02/25/23 11:01:17.96 +{"msg":"PASSED [sig-storage] Projected downwardAPI should update labels on modification [NodeConformance] [Conformance]","completed":225,"skipped":4146,"failed":0} ------------------------------ -• [0.682 seconds] -[sig-api-machinery] Discovery -test/e2e/apimachinery/framework.go:23 - should validate PreferredVersion for each APIGroup [Conformance] - test/e2e/apimachinery/discovery.go:122 +• [SLOW TEST] [6.912 seconds] +[sig-storage] Projected downwardAPI +test/e2e/common/storage/framework.go:23 + should update labels on modification [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:129 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Discovery + [BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:42:21.034 - Dec 14 16:42:21.034: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename discovery 12/14/22 16:42:21.036 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:42:21.058 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:42:21.064 - [BeforeEach] [sig-api-machinery] Discovery - test/e2e/apimachinery/discovery.go:43 - STEP: Setting up server cert 12/14/22 16:42:21.067 - [It] should validate PreferredVersion for each APIGroup [Conformance] - test/e2e/apimachinery/discovery.go:122 - Dec 14 16:42:21.674: INFO: Checking APIGroup: apiregistration.k8s.io - Dec 14 16:42:21.676: INFO: PreferredVersion.GroupVersion: apiregistration.k8s.io/v1 - Dec 14 16:42:21.676: INFO: Versions found [{apiregistration.k8s.io/v1 v1}] - Dec 14 16:42:21.676: INFO: apiregistration.k8s.io/v1 matches apiregistration.k8s.io/v1 - Dec 14 16:42:21.676: INFO: Checking APIGroup: apps - Dec 14 16:42:21.678: INFO: PreferredVersion.GroupVersion: apps/v1 - Dec 14 16:42:21.678: INFO: Versions found [{apps/v1 v1}] - Dec 14 16:42:21.678: INFO: apps/v1 matches apps/v1 - Dec 14 16:42:21.678: INFO: Checking APIGroup: events.k8s.io - Dec 14 16:42:21.679: INFO: PreferredVersion.GroupVersion: events.k8s.io/v1 - Dec 14 16:42:21.679: INFO: Versions found [{events.k8s.io/v1 v1}] - Dec 14 16:42:21.679: INFO: events.k8s.io/v1 matches events.k8s.io/v1 - Dec 14 16:42:21.679: INFO: Checking APIGroup: authentication.k8s.io - Dec 14 16:42:21.681: INFO: PreferredVersion.GroupVersion: authentication.k8s.io/v1 - Dec 14 16:42:21.681: INFO: Versions found [{authentication.k8s.io/v1 v1}] - Dec 14 16:42:21.681: INFO: authentication.k8s.io/v1 matches authentication.k8s.io/v1 - Dec 14 16:42:21.681: INFO: Checking APIGroup: authorization.k8s.io - Dec 14 16:42:21.682: INFO: PreferredVersion.GroupVersion: authorization.k8s.io/v1 - Dec 14 16:42:21.682: INFO: Versions found [{authorization.k8s.io/v1 v1}] - Dec 14 16:42:21.682: INFO: authorization.k8s.io/v1 matches authorization.k8s.io/v1 - Dec 14 16:42:21.682: INFO: Checking APIGroup: autoscaling - Dec 14 16:42:21.684: INFO: PreferredVersion.GroupVersion: autoscaling/v2 - Dec 14 16:42:21.684: INFO: Versions found [{autoscaling/v2 v2} {autoscaling/v1 v1} {autoscaling/v2beta2 v2beta2}] - Dec 14 16:42:21.684: INFO: autoscaling/v2 matches autoscaling/v2 - Dec 14 16:42:21.684: INFO: Checking APIGroup: batch - Dec 14 16:42:21.685: INFO: PreferredVersion.GroupVersion: batch/v1 - Dec 14 16:42:21.685: INFO: Versions found [{batch/v1 v1}] - Dec 14 16:42:21.685: INFO: batch/v1 matches batch/v1 - Dec 14 16:42:21.685: INFO: Checking APIGroup: certificates.k8s.io - Dec 14 16:42:21.687: INFO: PreferredVersion.GroupVersion: certificates.k8s.io/v1 - Dec 14 16:42:21.687: INFO: Versions found [{certificates.k8s.io/v1 v1}] - Dec 14 16:42:21.687: INFO: certificates.k8s.io/v1 matches certificates.k8s.io/v1 - Dec 14 16:42:21.687: INFO: Checking APIGroup: networking.k8s.io - Dec 14 16:42:21.688: INFO: PreferredVersion.GroupVersion: networking.k8s.io/v1 - Dec 14 16:42:21.688: INFO: Versions found [{networking.k8s.io/v1 v1}] - Dec 14 16:42:21.688: INFO: networking.k8s.io/v1 matches networking.k8s.io/v1 - Dec 14 16:42:21.688: INFO: Checking APIGroup: policy - Dec 14 16:42:21.690: INFO: PreferredVersion.GroupVersion: policy/v1 - Dec 14 16:42:21.690: INFO: Versions found [{policy/v1 v1}] - Dec 14 16:42:21.690: INFO: policy/v1 matches policy/v1 - Dec 14 16:42:21.690: INFO: Checking APIGroup: rbac.authorization.k8s.io - Dec 14 16:42:21.691: INFO: PreferredVersion.GroupVersion: rbac.authorization.k8s.io/v1 - Dec 14 16:42:21.691: INFO: Versions found [{rbac.authorization.k8s.io/v1 v1}] - Dec 14 16:42:21.691: INFO: rbac.authorization.k8s.io/v1 matches rbac.authorization.k8s.io/v1 - Dec 14 16:42:21.691: INFO: Checking APIGroup: storage.k8s.io - Dec 14 16:42:21.692: INFO: PreferredVersion.GroupVersion: storage.k8s.io/v1 - Dec 14 16:42:21.692: INFO: Versions found [{storage.k8s.io/v1 v1} {storage.k8s.io/v1beta1 v1beta1}] - Dec 14 16:42:21.692: INFO: storage.k8s.io/v1 matches storage.k8s.io/v1 - Dec 14 16:42:21.692: INFO: Checking APIGroup: admissionregistration.k8s.io - Dec 14 16:42:21.694: INFO: PreferredVersion.GroupVersion: admissionregistration.k8s.io/v1 - Dec 14 16:42:21.694: INFO: Versions found [{admissionregistration.k8s.io/v1 v1}] - Dec 14 16:42:21.694: INFO: admissionregistration.k8s.io/v1 matches admissionregistration.k8s.io/v1 - Dec 14 16:42:21.694: INFO: Checking APIGroup: apiextensions.k8s.io - Dec 14 16:42:21.695: INFO: PreferredVersion.GroupVersion: apiextensions.k8s.io/v1 - Dec 14 16:42:21.695: INFO: Versions found [{apiextensions.k8s.io/v1 v1}] - Dec 14 16:42:21.695: INFO: apiextensions.k8s.io/v1 matches apiextensions.k8s.io/v1 - Dec 14 16:42:21.695: INFO: Checking APIGroup: scheduling.k8s.io - Dec 14 16:42:21.697: INFO: PreferredVersion.GroupVersion: scheduling.k8s.io/v1 - Dec 14 16:42:21.697: INFO: Versions found [{scheduling.k8s.io/v1 v1}] - Dec 14 16:42:21.697: INFO: scheduling.k8s.io/v1 matches scheduling.k8s.io/v1 - Dec 14 16:42:21.697: INFO: Checking APIGroup: coordination.k8s.io - Dec 14 16:42:21.698: INFO: PreferredVersion.GroupVersion: coordination.k8s.io/v1 - Dec 14 16:42:21.698: INFO: Versions found [{coordination.k8s.io/v1 v1}] - Dec 14 16:42:21.698: INFO: coordination.k8s.io/v1 matches coordination.k8s.io/v1 - Dec 14 16:42:21.698: INFO: Checking APIGroup: node.k8s.io - Dec 14 16:42:21.699: INFO: PreferredVersion.GroupVersion: node.k8s.io/v1 - Dec 14 16:42:21.699: INFO: Versions found [{node.k8s.io/v1 v1}] - Dec 14 16:42:21.699: INFO: node.k8s.io/v1 matches node.k8s.io/v1 - Dec 14 16:42:21.699: INFO: Checking APIGroup: discovery.k8s.io - Dec 14 16:42:21.700: INFO: PreferredVersion.GroupVersion: discovery.k8s.io/v1 - Dec 14 16:42:21.700: INFO: Versions found [{discovery.k8s.io/v1 v1}] - Dec 14 16:42:21.700: INFO: discovery.k8s.io/v1 matches discovery.k8s.io/v1 - Dec 14 16:42:21.700: INFO: Checking APIGroup: flowcontrol.apiserver.k8s.io - Dec 14 16:42:21.702: INFO: PreferredVersion.GroupVersion: flowcontrol.apiserver.k8s.io/v1beta2 - Dec 14 16:42:21.702: INFO: Versions found [{flowcontrol.apiserver.k8s.io/v1beta2 v1beta2} {flowcontrol.apiserver.k8s.io/v1beta1 v1beta1}] - Dec 14 16:42:21.702: INFO: flowcontrol.apiserver.k8s.io/v1beta2 matches flowcontrol.apiserver.k8s.io/v1beta2 - [AfterEach] [sig-api-machinery] Discovery + STEP: Creating a kubernetes client 02/25/23 11:01:11.064 + Feb 25 11:01:11.064: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 11:01:11.07 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:01:11.103 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:01:11.109 + [BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 + [It] should update labels on modification [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:129 + STEP: Creating the pod 02/25/23 11:01:11.115 + Feb 25 11:01:11.133: INFO: Waiting up to 5m0s for pod "labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9" in namespace "projected-6214" to be "running and ready" + Feb 25 11:01:11.143: INFO: Pod "labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9": Phase="Pending", Reason="", readiness=false. Elapsed: 9.892354ms + Feb 25 11:01:11.143: INFO: The phase of Pod labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:01:13.164: INFO: Pod "labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031280921s + Feb 25 11:01:13.164: INFO: The phase of Pod labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:01:15.151: INFO: Pod "labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9": Phase="Running", Reason="", readiness=true. Elapsed: 4.017752184s + Feb 25 11:01:15.151: INFO: The phase of Pod labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9 is Running (Ready = true) + Feb 25 11:01:15.151: INFO: Pod "labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9" satisfied condition "running and ready" + Feb 25 11:01:15.699: INFO: Successfully updated pod "labelsupdatedf2dac3e-0296-423c-96f1-34f4089adea9" + [AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 - Dec 14 16:42:21.702: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "discovery-9662" for this suite. 12/14/22 16:42:21.709 + Feb 25 11:01:17.947: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-6214" for this suite. 02/25/23 11:01:17.96 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - updates the published spec when one version gets renamed [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:390 -[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +[sig-network] Networking Granular Checks: Pods + should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:122 +[BeforeEach] [sig-network] Networking test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:42:21.717 -Dec 14 16:42:21.717: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 16:42:21.72 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:42:21.748 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:42:21.756 -[It] updates the published spec when one version gets renamed [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:390 -STEP: set up a multi version CRD 12/14/22 16:42:21.778 -Dec 14 16:42:21.780: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: rename a version 12/14/22 16:42:31.706 -STEP: check the new version name is served 12/14/22 16:42:31.734 -STEP: check the old version name is removed 12/14/22 16:42:35.919 -STEP: check the other version is not changed 12/14/22 16:42:37.683 -[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 11:01:17.983 +Feb 25 11:01:17.983: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename pod-network-test 02/25/23 11:01:17.985 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:01:18.024 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:01:18.034 +[It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:122 +STEP: Performing setup for networking test in namespace pod-network-test-5290 02/25/23 11:01:18.043 +STEP: creating a selector 02/25/23 11:01:18.043 +STEP: Creating the service pods in kubernetes 02/25/23 11:01:18.044 +Feb 25 11:01:18.044: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable +Feb 25 11:01:18.197: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-5290" to be "running and ready" +Feb 25 11:01:18.242: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 44.311499ms +Feb 25 11:01:18.242: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:01:20.257: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.059141004s +Feb 25 11:01:20.257: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:01:22.254: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.056404278s +Feb 25 11:01:22.254: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:01:24.256: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.058075202s +Feb 25 11:01:24.256: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:01:26.251: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.053068312s +Feb 25 11:01:26.251: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:01:28.253: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.055571574s +Feb 25 11:01:28.253: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:01:30.250: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 12.05212736s +Feb 25 11:01:30.250: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:01:32.252: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 14.054930218s +Feb 25 11:01:32.253: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:01:34.255: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 16.057367986s +Feb 25 11:01:34.255: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:01:36.252: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 18.054350443s +Feb 25 11:01:36.252: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:01:38.252: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 20.054056699s +Feb 25 11:01:38.252: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:01:40.256: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 22.058433069s +Feb 25 11:01:40.256: INFO: The phase of Pod netserver-0 is Running (Ready = true) +Feb 25 11:01:40.256: INFO: Pod "netserver-0" satisfied condition "running and ready" +Feb 25 11:01:40.264: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-5290" to be "running and ready" +Feb 25 11:01:40.281: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 17.147364ms +Feb 25 11:01:40.281: INFO: The phase of Pod netserver-1 is Running (Ready = true) +Feb 25 11:01:40.281: INFO: Pod "netserver-1" satisfied condition "running and ready" +Feb 25 11:01:40.289: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-5290" to be "running and ready" +Feb 25 11:01:40.296: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 6.720393ms +Feb 25 11:01:40.296: INFO: The phase of Pod netserver-2 is Running (Ready = true) +Feb 25 11:01:40.296: INFO: Pod "netserver-2" satisfied condition "running and ready" +STEP: Creating test pods 02/25/23 11:01:40.305 +Feb 25 11:01:40.347: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-5290" to be "running" +Feb 25 11:01:40.364: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 16.837132ms +Feb 25 11:01:42.374: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02661995s +Feb 25 11:01:44.375: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.028078965s +Feb 25 11:01:44.375: INFO: Pod "test-container-pod" satisfied condition "running" +Feb 25 11:01:44.389: INFO: Waiting up to 5m0s for pod "host-test-container-pod" in namespace "pod-network-test-5290" to be "running" +Feb 25 11:01:44.397: INFO: Pod "host-test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 7.496292ms +Feb 25 11:01:44.397: INFO: Pod "host-test-container-pod" satisfied condition "running" +Feb 25 11:01:44.404: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 +Feb 25 11:01:44.404: INFO: Going to poll 10.233.65.6 on port 8081 at least 0 times, with a maximum of 39 tries before failing +Feb 25 11:01:44.415: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.233.65.6 8081 | grep -v '^\s*$'] Namespace:pod-network-test-5290 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 11:01:44.415: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 11:01:44.426: INFO: ExecWithOptions: Clientset creation +Feb 25 11:01:44.427: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-5290/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostName+%7C+nc+-w+1+-u+10.233.65.6+8081+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) +Feb 25 11:01:45.597: INFO: Found all 1 expected endpoints: [netserver-0] +Feb 25 11:01:45.597: INFO: Going to poll 10.233.64.231 on port 8081 at least 0 times, with a maximum of 39 tries before failing +Feb 25 11:01:45.605: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.233.64.231 8081 | grep -v '^\s*$'] Namespace:pod-network-test-5290 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 11:01:45.605: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 11:01:45.608: INFO: ExecWithOptions: Clientset creation +Feb 25 11:01:45.608: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-5290/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostName+%7C+nc+-w+1+-u+10.233.64.231+8081+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) +Feb 25 11:01:47.295: INFO: Found all 1 expected endpoints: [netserver-1] +Feb 25 11:01:47.295: INFO: Going to poll 10.233.66.94 on port 8081 at least 0 times, with a maximum of 39 tries before failing +Feb 25 11:01:47.304: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.233.66.94 8081 | grep -v '^\s*$'] Namespace:pod-network-test-5290 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 11:01:47.304: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 11:01:47.306: INFO: ExecWithOptions: Clientset creation +Feb 25 11:01:47.306: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-5290/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostName+%7C+nc+-w+1+-u+10.233.66.94+8081+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) +Feb 25 11:01:48.481: INFO: Found all 1 expected endpoints: [netserver-2] +[AfterEach] [sig-network] Networking test/e2e/framework/framework.go:187 -Dec 14 16:42:46.213: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-publish-openapi-3136" for this suite. 12/14/22 16:42:46.232 -{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] updates the published spec when one version gets renamed [Conformance]","completed":230,"skipped":4317,"failed":0} +Feb 25 11:01:48.481: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pod-network-test-5290" for this suite. 02/25/23 11:01:48.495 +{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance]","completed":226,"skipped":4191,"failed":0} ------------------------------ -• [SLOW TEST] [24.526 seconds] -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - updates the published spec when one version gets renamed [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:390 +• [SLOW TEST] [30.531 seconds] +[sig-network] Networking +test/e2e/common/network/framework.go:23 + Granular Checks: Pods + test/e2e/common/network/networking.go:32 + should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:122 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + [BeforeEach] [sig-network] Networking test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:42:21.717 - Dec 14 16:42:21.717: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 16:42:21.72 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:42:21.748 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:42:21.756 - [It] updates the published spec when one version gets renamed [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:390 - STEP: set up a multi version CRD 12/14/22 16:42:21.778 - Dec 14 16:42:21.780: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: rename a version 12/14/22 16:42:31.706 - STEP: check the new version name is served 12/14/22 16:42:31.734 - STEP: check the old version name is removed 12/14/22 16:42:35.919 - STEP: check the other version is not changed 12/14/22 16:42:37.683 - [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 11:01:17.983 + Feb 25 11:01:17.983: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename pod-network-test 02/25/23 11:01:17.985 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:01:18.024 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:01:18.034 + [It] should function for node-pod communication: udp [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:122 + STEP: Performing setup for networking test in namespace pod-network-test-5290 02/25/23 11:01:18.043 + STEP: creating a selector 02/25/23 11:01:18.043 + STEP: Creating the service pods in kubernetes 02/25/23 11:01:18.044 + Feb 25 11:01:18.044: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable + Feb 25 11:01:18.197: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-5290" to be "running and ready" + Feb 25 11:01:18.242: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 44.311499ms + Feb 25 11:01:18.242: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:01:20.257: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.059141004s + Feb 25 11:01:20.257: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:01:22.254: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.056404278s + Feb 25 11:01:22.254: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:01:24.256: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.058075202s + Feb 25 11:01:24.256: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:01:26.251: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.053068312s + Feb 25 11:01:26.251: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:01:28.253: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.055571574s + Feb 25 11:01:28.253: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:01:30.250: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 12.05212736s + Feb 25 11:01:30.250: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:01:32.252: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 14.054930218s + Feb 25 11:01:32.253: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:01:34.255: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 16.057367986s + Feb 25 11:01:34.255: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:01:36.252: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 18.054350443s + Feb 25 11:01:36.252: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:01:38.252: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 20.054056699s + Feb 25 11:01:38.252: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:01:40.256: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 22.058433069s + Feb 25 11:01:40.256: INFO: The phase of Pod netserver-0 is Running (Ready = true) + Feb 25 11:01:40.256: INFO: Pod "netserver-0" satisfied condition "running and ready" + Feb 25 11:01:40.264: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-5290" to be "running and ready" + Feb 25 11:01:40.281: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 17.147364ms + Feb 25 11:01:40.281: INFO: The phase of Pod netserver-1 is Running (Ready = true) + Feb 25 11:01:40.281: INFO: Pod "netserver-1" satisfied condition "running and ready" + Feb 25 11:01:40.289: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-5290" to be "running and ready" + Feb 25 11:01:40.296: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 6.720393ms + Feb 25 11:01:40.296: INFO: The phase of Pod netserver-2 is Running (Ready = true) + Feb 25 11:01:40.296: INFO: Pod "netserver-2" satisfied condition "running and ready" + STEP: Creating test pods 02/25/23 11:01:40.305 + Feb 25 11:01:40.347: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-5290" to be "running" + Feb 25 11:01:40.364: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 16.837132ms + Feb 25 11:01:42.374: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02661995s + Feb 25 11:01:44.375: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.028078965s + Feb 25 11:01:44.375: INFO: Pod "test-container-pod" satisfied condition "running" + Feb 25 11:01:44.389: INFO: Waiting up to 5m0s for pod "host-test-container-pod" in namespace "pod-network-test-5290" to be "running" + Feb 25 11:01:44.397: INFO: Pod "host-test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 7.496292ms + Feb 25 11:01:44.397: INFO: Pod "host-test-container-pod" satisfied condition "running" + Feb 25 11:01:44.404: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 + Feb 25 11:01:44.404: INFO: Going to poll 10.233.65.6 on port 8081 at least 0 times, with a maximum of 39 tries before failing + Feb 25 11:01:44.415: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.233.65.6 8081 | grep -v '^\s*$'] Namespace:pod-network-test-5290 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 11:01:44.415: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 11:01:44.426: INFO: ExecWithOptions: Clientset creation + Feb 25 11:01:44.427: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-5290/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostName+%7C+nc+-w+1+-u+10.233.65.6+8081+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) + Feb 25 11:01:45.597: INFO: Found all 1 expected endpoints: [netserver-0] + Feb 25 11:01:45.597: INFO: Going to poll 10.233.64.231 on port 8081 at least 0 times, with a maximum of 39 tries before failing + Feb 25 11:01:45.605: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.233.64.231 8081 | grep -v '^\s*$'] Namespace:pod-network-test-5290 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 11:01:45.605: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 11:01:45.608: INFO: ExecWithOptions: Clientset creation + Feb 25 11:01:45.608: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-5290/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostName+%7C+nc+-w+1+-u+10.233.64.231+8081+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) + Feb 25 11:01:47.295: INFO: Found all 1 expected endpoints: [netserver-1] + Feb 25 11:01:47.295: INFO: Going to poll 10.233.66.94 on port 8081 at least 0 times, with a maximum of 39 tries before failing + Feb 25 11:01:47.304: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostName | nc -w 1 -u 10.233.66.94 8081 | grep -v '^\s*$'] Namespace:pod-network-test-5290 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 11:01:47.304: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 11:01:47.306: INFO: ExecWithOptions: Clientset creation + Feb 25 11:01:47.306: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-5290/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostName+%7C+nc+-w+1+-u+10.233.66.94+8081+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) + Feb 25 11:01:48.481: INFO: Found all 1 expected endpoints: [netserver-2] + [AfterEach] [sig-network] Networking test/e2e/framework/framework.go:187 - Dec 14 16:42:46.213: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "crd-publish-openapi-3136" for this suite. 12/14/22 16:42:46.232 + Feb 25 11:01:48.481: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "pod-network-test-5290" for this suite. 02/25/23 11:01:48.495 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSS +SSS ------------------------------ -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - should mutate custom resource with different stored version [Conformance] - test/e2e/apimachinery/webhook.go:322 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +[sig-node] Probing container + should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:180 +[BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:42:46.245 -Dec 14 16:42:46.245: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 16:42:46.251 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:42:46.277 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:42:46.281 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 16:42:46.305 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:42:47.548 -STEP: Deploying the webhook pod 12/14/22 16:42:47.565 -STEP: Wait for the deployment to be ready 12/14/22 16:42:47.588 -Dec 14 16:42:47.605: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set -Dec 14 16:42:49.629: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 42, 47, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 42, 47, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 42, 47, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 42, 47, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} -STEP: Deploying the webhook service 12/14/22 16:42:51.637 -STEP: Verifying the service has paired with the endpoint 12/14/22 16:42:51.656 -Dec 14 16:42:52.656: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] should mutate custom resource with different stored version [Conformance] - test/e2e/apimachinery/webhook.go:322 -Dec 14 16:42:52.661: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Registering the mutating webhook for custom resource e2e-test-webhook-6821-crds.webhook.example.com via the AdmissionRegistration API 12/14/22 16:42:53.189 -STEP: Creating a custom resource while v1 is storage version 12/14/22 16:42:53.22 -STEP: Patching Custom Resource Definition to set v2 as storage 12/14/22 16:42:55.389 -STEP: Patching the custom resource while v2 is storage version 12/14/22 16:42:55.413 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 11:01:48.518 +Feb 25 11:01:48.518: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-probe 02/25/23 11:01:48.523 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:01:48.566 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:01:48.577 +[BeforeEach] [sig-node] Probing container + test/e2e/common/node/container_probe.go:59 +[It] should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:180 +STEP: Creating pod liveness-3a702bed-4bb6-48c1-bd21-47be6e644136 in namespace container-probe-5320 02/25/23 11:01:48.585 +Feb 25 11:01:48.607: INFO: Waiting up to 5m0s for pod "liveness-3a702bed-4bb6-48c1-bd21-47be6e644136" in namespace "container-probe-5320" to be "not pending" +Feb 25 11:01:48.615: INFO: Pod "liveness-3a702bed-4bb6-48c1-bd21-47be6e644136": Phase="Pending", Reason="", readiness=false. Elapsed: 7.702801ms +Feb 25 11:01:50.627: INFO: Pod "liveness-3a702bed-4bb6-48c1-bd21-47be6e644136": Phase="Running", Reason="", readiness=true. Elapsed: 2.019460988s +Feb 25 11:01:50.627: INFO: Pod "liveness-3a702bed-4bb6-48c1-bd21-47be6e644136" satisfied condition "not pending" +Feb 25 11:01:50.627: INFO: Started pod liveness-3a702bed-4bb6-48c1-bd21-47be6e644136 in namespace container-probe-5320 +STEP: checking the pod's current state and verifying that restartCount is present 02/25/23 11:01:50.627 +Feb 25 11:01:50.635: INFO: Initial restart count of pod liveness-3a702bed-4bb6-48c1-bd21-47be6e644136 is 0 +STEP: deleting the pod 02/25/23 11:05:52.185 +[AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 -Dec 14 16:42:55.967: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-6695" for this suite. 12/14/22 16:42:55.974 -STEP: Destroying namespace "webhook-6695-markers" for this suite. 12/14/22 16:42:55.985 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource with different stored version [Conformance]","completed":231,"skipped":4329,"failed":0} +Feb 25 11:05:52.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-5320" for this suite. 02/25/23 11:05:52.25 +{"msg":"PASSED [sig-node] Probing container should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance]","completed":227,"skipped":4194,"failed":0} ------------------------------ -• [SLOW TEST] [9.945 seconds] -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - should mutate custom resource with different stored version [Conformance] - test/e2e/apimachinery/webhook.go:322 +• [SLOW TEST] [243.757 seconds] +[sig-node] Probing container +test/e2e/common/node/framework.go:23 + should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:180 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + [BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:42:46.245 - Dec 14 16:42:46.245: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 16:42:46.251 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:42:46.277 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:42:46.281 - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 16:42:46.305 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:42:47.548 - STEP: Deploying the webhook pod 12/14/22 16:42:47.565 - STEP: Wait for the deployment to be ready 12/14/22 16:42:47.588 - Dec 14 16:42:47.605: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set - Dec 14 16:42:49.629: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 42, 47, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 42, 47, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 42, 47, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 42, 47, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} - STEP: Deploying the webhook service 12/14/22 16:42:51.637 - STEP: Verifying the service has paired with the endpoint 12/14/22 16:42:51.656 - Dec 14 16:42:52.656: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] should mutate custom resource with different stored version [Conformance] - test/e2e/apimachinery/webhook.go:322 - Dec 14 16:42:52.661: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Registering the mutating webhook for custom resource e2e-test-webhook-6821-crds.webhook.example.com via the AdmissionRegistration API 12/14/22 16:42:53.189 - STEP: Creating a custom resource while v1 is storage version 12/14/22 16:42:53.22 - STEP: Patching Custom Resource Definition to set v2 as storage 12/14/22 16:42:55.389 - STEP: Patching the custom resource while v2 is storage version 12/14/22 16:42:55.413 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 11:01:48.518 + Feb 25 11:01:48.518: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-probe 02/25/23 11:01:48.523 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:01:48.566 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:01:48.577 + [BeforeEach] [sig-node] Probing container + test/e2e/common/node/container_probe.go:59 + [It] should *not* be restarted with a tcp:8080 liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:180 + STEP: Creating pod liveness-3a702bed-4bb6-48c1-bd21-47be6e644136 in namespace container-probe-5320 02/25/23 11:01:48.585 + Feb 25 11:01:48.607: INFO: Waiting up to 5m0s for pod "liveness-3a702bed-4bb6-48c1-bd21-47be6e644136" in namespace "container-probe-5320" to be "not pending" + Feb 25 11:01:48.615: INFO: Pod "liveness-3a702bed-4bb6-48c1-bd21-47be6e644136": Phase="Pending", Reason="", readiness=false. Elapsed: 7.702801ms + Feb 25 11:01:50.627: INFO: Pod "liveness-3a702bed-4bb6-48c1-bd21-47be6e644136": Phase="Running", Reason="", readiness=true. Elapsed: 2.019460988s + Feb 25 11:01:50.627: INFO: Pod "liveness-3a702bed-4bb6-48c1-bd21-47be6e644136" satisfied condition "not pending" + Feb 25 11:01:50.627: INFO: Started pod liveness-3a702bed-4bb6-48c1-bd21-47be6e644136 in namespace container-probe-5320 + STEP: checking the pod's current state and verifying that restartCount is present 02/25/23 11:01:50.627 + Feb 25 11:01:50.635: INFO: Initial restart count of pod liveness-3a702bed-4bb6-48c1-bd21-47be6e644136 is 0 + STEP: deleting the pod 02/25/23 11:05:52.185 + [AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 - Dec 14 16:42:55.967: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-6695" for this suite. 12/14/22 16:42:55.974 - STEP: Destroying namespace "webhook-6695-markers" for this suite. 12/14/22 16:42:55.985 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 + Feb 25 11:05:52.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-probe-5320" for this suite. 02/25/23 11:05:52.25 << End Captured GinkgoWriter Output ------------------------------ -S +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] - test/e2e/apimachinery/webhook.go:276 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +[sig-apps] CronJob + should not schedule jobs when suspended [Slow] [Conformance] + test/e2e/apps/cronjob.go:96 +[BeforeEach] [sig-apps] CronJob test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:42:56.192 -Dec 14 16:42:56.192: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 16:42:56.194 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:42:56.242 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:42:56.246 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 16:42:56.331 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:42:57.025 -STEP: Deploying the webhook pod 12/14/22 16:42:57.031 -STEP: Wait for the deployment to be ready 12/14/22 16:42:57.048 -Dec 14 16:42:57.059: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created -STEP: Deploying the webhook service 12/14/22 16:42:59.08 -STEP: Verifying the service has paired with the endpoint 12/14/22 16:42:59.102 -Dec 14 16:43:00.102: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] - test/e2e/apimachinery/webhook.go:276 -STEP: Registering a validating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API 12/14/22 16:43:00.108 -STEP: Registering a mutating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API 12/14/22 16:43:00.13 -STEP: Creating a dummy validating-webhook-configuration object 12/14/22 16:43:00.16 -STEP: Deleting the validating-webhook-configuration, which should be possible to remove 12/14/22 16:43:00.172 -STEP: Creating a dummy mutating-webhook-configuration object 12/14/22 16:43:00.18 -STEP: Deleting the mutating-webhook-configuration, which should be possible to remove 12/14/22 16:43:00.192 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 11:05:52.294 +Feb 25 11:05:52.294: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename cronjob 02/25/23 11:05:52.305 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:05:52.339 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:05:52.346 +[It] should not schedule jobs when suspended [Slow] [Conformance] + test/e2e/apps/cronjob.go:96 +STEP: Creating a suspended cronjob 02/25/23 11:05:52.354 +STEP: Ensuring no jobs are scheduled 02/25/23 11:05:52.371 +STEP: Ensuring no job exists by listing jobs explicitly 02/25/23 11:10:52.409 +STEP: Removing cronjob 02/25/23 11:10:52.422 +[AfterEach] [sig-apps] CronJob test/e2e/framework/framework.go:187 -Dec 14 16:43:00.258: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-9479" for this suite. 12/14/22 16:43:00.27 -STEP: Destroying namespace "webhook-9479-markers" for this suite. 12/14/22 16:43:00.299 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance]","completed":232,"skipped":4330,"failed":0} +Feb 25 11:10:52.445: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "cronjob-4690" for this suite. 02/25/23 11:10:52.455 +{"msg":"PASSED [sig-apps] CronJob should not schedule jobs when suspended [Slow] [Conformance]","completed":228,"skipped":4242,"failed":0} ------------------------------ -• [4.235 seconds] -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] - test/e2e/apimachinery/webhook.go:276 +• [SLOW TEST] [300.177 seconds] +[sig-apps] CronJob +test/e2e/apps/framework.go:23 + should not schedule jobs when suspended [Slow] [Conformance] + test/e2e/apps/cronjob.go:96 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + [BeforeEach] [sig-apps] CronJob test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:42:56.192 - Dec 14 16:42:56.192: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 16:42:56.194 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:42:56.242 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:42:56.246 - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 16:42:56.331 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:42:57.025 - STEP: Deploying the webhook pod 12/14/22 16:42:57.031 - STEP: Wait for the deployment to be ready 12/14/22 16:42:57.048 - Dec 14 16:42:57.059: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created - STEP: Deploying the webhook service 12/14/22 16:42:59.08 - STEP: Verifying the service has paired with the endpoint 12/14/22 16:42:59.102 - Dec 14 16:43:00.102: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] should not be able to mutate or prevent deletion of webhook configuration objects [Conformance] - test/e2e/apimachinery/webhook.go:276 - STEP: Registering a validating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API 12/14/22 16:43:00.108 - STEP: Registering a mutating webhook on ValidatingWebhookConfiguration and MutatingWebhookConfiguration objects, via the AdmissionRegistration API 12/14/22 16:43:00.13 - STEP: Creating a dummy validating-webhook-configuration object 12/14/22 16:43:00.16 - STEP: Deleting the validating-webhook-configuration, which should be possible to remove 12/14/22 16:43:00.172 - STEP: Creating a dummy mutating-webhook-configuration object 12/14/22 16:43:00.18 - STEP: Deleting the mutating-webhook-configuration, which should be possible to remove 12/14/22 16:43:00.192 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 11:05:52.294 + Feb 25 11:05:52.294: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename cronjob 02/25/23 11:05:52.305 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:05:52.339 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:05:52.346 + [It] should not schedule jobs when suspended [Slow] [Conformance] + test/e2e/apps/cronjob.go:96 + STEP: Creating a suspended cronjob 02/25/23 11:05:52.354 + STEP: Ensuring no jobs are scheduled 02/25/23 11:05:52.371 + STEP: Ensuring no job exists by listing jobs explicitly 02/25/23 11:10:52.409 + STEP: Removing cronjob 02/25/23 11:10:52.422 + [AfterEach] [sig-apps] CronJob test/e2e/framework/framework.go:187 - Dec 14 16:43:00.258: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-9479" for this suite. 12/14/22 16:43:00.27 - STEP: Destroying namespace "webhook-9479-markers" for this suite. 12/14/22 16:43:00.299 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 + Feb 25 11:10:52.445: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "cronjob-4690" for this suite. 02/25/23 11:10:52.455 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSSSSSS ------------------------------ -[sig-scheduling] LimitRange - should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] - test/e2e/scheduling/limit_range.go:57 -[BeforeEach] [sig-scheduling] LimitRange +[sig-storage] ConfigMap + binary data should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:174 +[BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:43:00.431 -Dec 14 16:43:00.431: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename limitrange 12/14/22 16:43:00.433 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:43:00.507 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:43:00.51 -[It] should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] - test/e2e/scheduling/limit_range.go:57 -STEP: Creating a LimitRange 12/14/22 16:43:00.514 -STEP: Setting up watch 12/14/22 16:43:00.515 -STEP: Submitting a LimitRange 12/14/22 16:43:00.621 -STEP: Verifying LimitRange creation was observed 12/14/22 16:43:00.632 -STEP: Fetching the LimitRange to ensure it has proper values 12/14/22 16:43:00.632 -Dec 14 16:43:00.638: INFO: Verifying requests: expected map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] with actual map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] -Dec 14 16:43:00.639: INFO: Verifying limits: expected map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] -STEP: Creating a Pod with no resource requirements 12/14/22 16:43:00.639 -STEP: Ensuring Pod has resource requirements applied from LimitRange 12/14/22 16:43:00.65 -Dec 14 16:43:00.655: INFO: Verifying requests: expected map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] with actual map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] -Dec 14 16:43:00.655: INFO: Verifying limits: expected map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] -STEP: Creating a Pod with partial resource requirements 12/14/22 16:43:00.655 -STEP: Ensuring Pod has merged resource requirements applied from LimitRange 12/14/22 16:43:00.663 -Dec 14 16:43:00.669: INFO: Verifying requests: expected map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{161061273600 0} {} 150Gi BinarySI} memory:{{157286400 0} {} 150Mi BinarySI}] with actual map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{161061273600 0} {} 150Gi BinarySI} memory:{{157286400 0} {} 150Mi BinarySI}] -Dec 14 16:43:00.669: INFO: Verifying limits: expected map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] -STEP: Failing to create a Pod with less than min resources 12/14/22 16:43:00.669 -STEP: Failing to create a Pod with more than max resources 12/14/22 16:43:00.672 -STEP: Updating a LimitRange 12/14/22 16:43:00.674 -STEP: Verifying LimitRange updating is effective 12/14/22 16:43:00.682 -STEP: Creating a Pod with less than former min resources 12/14/22 16:43:02.689 -STEP: Failing to create a Pod with more than max resources 12/14/22 16:43:02.697 -STEP: Deleting a LimitRange 12/14/22 16:43:02.7 -STEP: Verifying the LimitRange was deleted 12/14/22 16:43:02.709 -Dec 14 16:43:07.719: INFO: limitRange is already deleted -STEP: Creating a Pod with more than former max resources 12/14/22 16:43:07.72 -[AfterEach] [sig-scheduling] LimitRange +STEP: Creating a kubernetes client 02/25/23 11:10:52.499 +Feb 25 11:10:52.499: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename configmap 02/25/23 11:10:52.515 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:10:52.542 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:10:52.549 +[It] binary data should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:174 +STEP: Creating configMap with name configmap-test-upd-533a3bec-945b-4353-a10d-3870b8808ab7 02/25/23 11:10:52.57 +STEP: Creating the pod 02/25/23 11:10:52.579 +Feb 25 11:10:52.603: INFO: Waiting up to 5m0s for pod "pod-configmaps-78b033a5-8c23-403c-ad35-591445ee61f2" in namespace "configmap-2449" to be "running" +Feb 25 11:10:52.611: INFO: Pod "pod-configmaps-78b033a5-8c23-403c-ad35-591445ee61f2": Phase="Pending", Reason="", readiness=false. Elapsed: 7.892183ms +Feb 25 11:10:54.621: INFO: Pod "pod-configmaps-78b033a5-8c23-403c-ad35-591445ee61f2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01840903s +Feb 25 11:10:56.620: INFO: Pod "pod-configmaps-78b033a5-8c23-403c-ad35-591445ee61f2": Phase="Running", Reason="", readiness=false. Elapsed: 4.016571264s +Feb 25 11:10:56.620: INFO: Pod "pod-configmaps-78b033a5-8c23-403c-ad35-591445ee61f2" satisfied condition "running" +STEP: Waiting for pod with text data 02/25/23 11:10:56.62 +STEP: Waiting for pod with binary data 02/25/23 11:10:56.661 +[AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 -Dec 14 16:43:07.738: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "limitrange-4334" for this suite. 12/14/22 16:43:07.76 -{"msg":"PASSED [sig-scheduling] LimitRange should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance]","completed":233,"skipped":4333,"failed":0} +Feb 25 11:10:56.675: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-2449" for this suite. 02/25/23 11:10:56.689 +{"msg":"PASSED [sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance]","completed":229,"skipped":4252,"failed":0} ------------------------------ -• [SLOW TEST] [7.350 seconds] -[sig-scheduling] LimitRange -test/e2e/scheduling/framework.go:40 - should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] - test/e2e/scheduling/limit_range.go:57 +• [4.209 seconds] +[sig-storage] ConfigMap +test/e2e/common/storage/framework.go:23 + binary data should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:174 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-scheduling] LimitRange + [BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:43:00.431 - Dec 14 16:43:00.431: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename limitrange 12/14/22 16:43:00.433 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:43:00.507 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:43:00.51 - [It] should create a LimitRange with defaults and ensure pod has those defaults applied. [Conformance] - test/e2e/scheduling/limit_range.go:57 - STEP: Creating a LimitRange 12/14/22 16:43:00.514 - STEP: Setting up watch 12/14/22 16:43:00.515 - STEP: Submitting a LimitRange 12/14/22 16:43:00.621 - STEP: Verifying LimitRange creation was observed 12/14/22 16:43:00.632 - STEP: Fetching the LimitRange to ensure it has proper values 12/14/22 16:43:00.632 - Dec 14 16:43:00.638: INFO: Verifying requests: expected map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] with actual map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] - Dec 14 16:43:00.639: INFO: Verifying limits: expected map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] - STEP: Creating a Pod with no resource requirements 12/14/22 16:43:00.639 - STEP: Ensuring Pod has resource requirements applied from LimitRange 12/14/22 16:43:00.65 - Dec 14 16:43:00.655: INFO: Verifying requests: expected map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] with actual map[cpu:{{100 -3} {} 100m DecimalSI} ephemeral-storage:{{214748364800 0} {} BinarySI} memory:{{209715200 0} {} BinarySI}] - Dec 14 16:43:00.655: INFO: Verifying limits: expected map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{500 -3} {} 500m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] - STEP: Creating a Pod with partial resource requirements 12/14/22 16:43:00.655 - STEP: Ensuring Pod has merged resource requirements applied from LimitRange 12/14/22 16:43:00.663 - Dec 14 16:43:00.669: INFO: Verifying requests: expected map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{161061273600 0} {} 150Gi BinarySI} memory:{{157286400 0} {} 150Mi BinarySI}] with actual map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{161061273600 0} {} 150Gi BinarySI} memory:{{157286400 0} {} 150Mi BinarySI}] - Dec 14 16:43:00.669: INFO: Verifying limits: expected map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] with actual map[cpu:{{300 -3} {} 300m DecimalSI} ephemeral-storage:{{536870912000 0} {} 500Gi BinarySI} memory:{{524288000 0} {} 500Mi BinarySI}] - STEP: Failing to create a Pod with less than min resources 12/14/22 16:43:00.669 - STEP: Failing to create a Pod with more than max resources 12/14/22 16:43:00.672 - STEP: Updating a LimitRange 12/14/22 16:43:00.674 - STEP: Verifying LimitRange updating is effective 12/14/22 16:43:00.682 - STEP: Creating a Pod with less than former min resources 12/14/22 16:43:02.689 - STEP: Failing to create a Pod with more than max resources 12/14/22 16:43:02.697 - STEP: Deleting a LimitRange 12/14/22 16:43:02.7 - STEP: Verifying the LimitRange was deleted 12/14/22 16:43:02.709 - Dec 14 16:43:07.719: INFO: limitRange is already deleted - STEP: Creating a Pod with more than former max resources 12/14/22 16:43:07.72 - [AfterEach] [sig-scheduling] LimitRange + STEP: Creating a kubernetes client 02/25/23 11:10:52.499 + Feb 25 11:10:52.499: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename configmap 02/25/23 11:10:52.515 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:10:52.542 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:10:52.549 + [It] binary data should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:174 + STEP: Creating configMap with name configmap-test-upd-533a3bec-945b-4353-a10d-3870b8808ab7 02/25/23 11:10:52.57 + STEP: Creating the pod 02/25/23 11:10:52.579 + Feb 25 11:10:52.603: INFO: Waiting up to 5m0s for pod "pod-configmaps-78b033a5-8c23-403c-ad35-591445ee61f2" in namespace "configmap-2449" to be "running" + Feb 25 11:10:52.611: INFO: Pod "pod-configmaps-78b033a5-8c23-403c-ad35-591445ee61f2": Phase="Pending", Reason="", readiness=false. Elapsed: 7.892183ms + Feb 25 11:10:54.621: INFO: Pod "pod-configmaps-78b033a5-8c23-403c-ad35-591445ee61f2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01840903s + Feb 25 11:10:56.620: INFO: Pod "pod-configmaps-78b033a5-8c23-403c-ad35-591445ee61f2": Phase="Running", Reason="", readiness=false. Elapsed: 4.016571264s + Feb 25 11:10:56.620: INFO: Pod "pod-configmaps-78b033a5-8c23-403c-ad35-591445ee61f2" satisfied condition "running" + STEP: Waiting for pod with text data 02/25/23 11:10:56.62 + STEP: Waiting for pod with binary data 02/25/23 11:10:56.661 + [AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 - Dec 14 16:43:07.738: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "limitrange-4334" for this suite. 12/14/22 16:43:07.76 + Feb 25 11:10:56.675: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "configmap-2449" for this suite. 02/25/23 11:10:56.689 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSS ------------------------------ -[sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath - runs ReplicaSets to verify preemption running path [Conformance] - test/e2e/scheduling/preemption.go:543 -[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:43:07.789 -Dec 14 16:43:07.789: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename sched-preemption 12/14/22 16:43:07.791 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:43:07.828 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:43:07.832 -[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:92 -Dec 14 16:43:07.859: INFO: Waiting up to 1m0s for all nodes to be ready -Dec 14 16:44:07.892: INFO: Waiting for terminating namespaces to be deleted... -[BeforeEach] PreemptionExecutionPath +[sig-storage] Projected configMap + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:88 +[BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:44:07.896 -Dec 14 16:44:07.896: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename sched-preemption-path 12/14/22 16:44:07.899 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:44:07.927 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:44:07.93 -[BeforeEach] PreemptionExecutionPath - test/e2e/scheduling/preemption.go:496 -STEP: Finding an available node 12/14/22 16:44:07.933 -STEP: Trying to launch a pod without a label to get a node which can launch it. 12/14/22 16:44:07.933 -Dec 14 16:44:07.942: INFO: Waiting up to 1m0s for pod "without-label" in namespace "sched-preemption-path-7732" to be "running" -Dec 14 16:44:07.947: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 4.382744ms -Dec 14 16:44:09.952: INFO: Pod "without-label": Phase="Running", Reason="", readiness=true. Elapsed: 2.010247446s -Dec 14 16:44:09.953: INFO: Pod "without-label" satisfied condition "running" -STEP: Explicitly delete pod here to free the resource it takes. 12/14/22 16:44:09.957 -Dec 14 16:44:09.972: INFO: found a healthy node: kiedai4iakae-3 -[It] runs ReplicaSets to verify preemption running path [Conformance] - test/e2e/scheduling/preemption.go:543 -Dec 14 16:44:22.113: INFO: pods created so far: [1 1 1] -Dec 14 16:44:22.113: INFO: length of pods created so far: 3 -Dec 14 16:44:24.151: INFO: pods created so far: [2 2 1] -[AfterEach] PreemptionExecutionPath - test/e2e/framework/framework.go:187 -Dec 14 16:44:31.152: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "sched-preemption-path-7732" for this suite. 12/14/22 16:44:31.165 -[AfterEach] PreemptionExecutionPath - test/e2e/scheduling/preemption.go:470 -[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] +STEP: Creating a kubernetes client 02/25/23 11:10:56.713 +Feb 25 11:10:56.713: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 11:10:56.715 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:10:56.758 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:10:56.769 +[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:88 +STEP: Creating configMap with name projected-configmap-test-volume-map-4b859330-69b8-41b7-96e0-d919bf915a68 02/25/23 11:10:56.781 +STEP: Creating a pod to test consume configMaps 02/25/23 11:10:56.792 +Feb 25 11:10:56.823: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1" in namespace "projected-5843" to be "Succeeded or Failed" +Feb 25 11:10:56.835: INFO: Pod "pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1": Phase="Pending", Reason="", readiness=false. Elapsed: 11.946884ms +Feb 25 11:10:58.845: INFO: Pod "pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1": Phase="Running", Reason="", readiness=true. Elapsed: 2.022036129s +Feb 25 11:11:00.846: INFO: Pod "pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1": Phase="Running", Reason="", readiness=false. Elapsed: 4.022969986s +Feb 25 11:11:02.854: INFO: Pod "pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.030559788s +STEP: Saw pod success 02/25/23 11:11:02.854 +Feb 25 11:11:02.854: INFO: Pod "pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1" satisfied condition "Succeeded or Failed" +Feb 25 11:11:02.863: INFO: Trying to get logs from node diek3dua9gay-2 pod pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1 container agnhost-container: +STEP: delete the pod 02/25/23 11:11:02.974 +Feb 25 11:11:03.000: INFO: Waiting for pod pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1 to disappear +Feb 25 11:11:03.006: INFO: Pod pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1 no longer exists +[AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 -Dec 14 16:44:31.227: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "sched-preemption-1072" for this suite. 12/14/22 16:44:31.234 -[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:80 -{"msg":"PASSED [sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath runs ReplicaSets to verify preemption running path [Conformance]","completed":234,"skipped":4360,"failed":0} +Feb 25 11:11:03.007: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-5843" for this suite. 02/25/23 11:11:03.017 +{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","completed":230,"skipped":4267,"failed":0} ------------------------------ -• [SLOW TEST] [83.531 seconds] -[sig-scheduling] SchedulerPreemption [Serial] -test/e2e/scheduling/framework.go:40 - PreemptionExecutionPath - test/e2e/scheduling/preemption.go:458 - runs ReplicaSets to verify preemption running path [Conformance] - test/e2e/scheduling/preemption.go:543 +• [SLOW TEST] [6.320 seconds] +[sig-storage] Projected configMap +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:88 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:43:07.789 - Dec 14 16:43:07.789: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename sched-preemption 12/14/22 16:43:07.791 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:43:07.828 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:43:07.832 - [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:92 - Dec 14 16:43:07.859: INFO: Waiting up to 1m0s for all nodes to be ready - Dec 14 16:44:07.892: INFO: Waiting for terminating namespaces to be deleted... - [BeforeEach] PreemptionExecutionPath + [BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:44:07.896 - Dec 14 16:44:07.896: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename sched-preemption-path 12/14/22 16:44:07.899 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:44:07.927 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:44:07.93 - [BeforeEach] PreemptionExecutionPath - test/e2e/scheduling/preemption.go:496 - STEP: Finding an available node 12/14/22 16:44:07.933 - STEP: Trying to launch a pod without a label to get a node which can launch it. 12/14/22 16:44:07.933 - Dec 14 16:44:07.942: INFO: Waiting up to 1m0s for pod "without-label" in namespace "sched-preemption-path-7732" to be "running" - Dec 14 16:44:07.947: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 4.382744ms - Dec 14 16:44:09.952: INFO: Pod "without-label": Phase="Running", Reason="", readiness=true. Elapsed: 2.010247446s - Dec 14 16:44:09.953: INFO: Pod "without-label" satisfied condition "running" - STEP: Explicitly delete pod here to free the resource it takes. 12/14/22 16:44:09.957 - Dec 14 16:44:09.972: INFO: found a healthy node: kiedai4iakae-3 - [It] runs ReplicaSets to verify preemption running path [Conformance] - test/e2e/scheduling/preemption.go:543 - Dec 14 16:44:22.113: INFO: pods created so far: [1 1 1] - Dec 14 16:44:22.113: INFO: length of pods created so far: 3 - Dec 14 16:44:24.151: INFO: pods created so far: [2 2 1] - [AfterEach] PreemptionExecutionPath - test/e2e/framework/framework.go:187 - Dec 14 16:44:31.152: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "sched-preemption-path-7732" for this suite. 12/14/22 16:44:31.165 - [AfterEach] PreemptionExecutionPath - test/e2e/scheduling/preemption.go:470 - [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/framework/framework.go:187 - Dec 14 16:44:31.227: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "sched-preemption-1072" for this suite. 12/14/22 16:44:31.234 - [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:80 + STEP: Creating a kubernetes client 02/25/23 11:10:56.713 + Feb 25 11:10:56.713: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 11:10:56.715 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:10:56.758 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:10:56.769 + [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:88 + STEP: Creating configMap with name projected-configmap-test-volume-map-4b859330-69b8-41b7-96e0-d919bf915a68 02/25/23 11:10:56.781 + STEP: Creating a pod to test consume configMaps 02/25/23 11:10:56.792 + Feb 25 11:10:56.823: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1" in namespace "projected-5843" to be "Succeeded or Failed" + Feb 25 11:10:56.835: INFO: Pod "pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1": Phase="Pending", Reason="", readiness=false. Elapsed: 11.946884ms + Feb 25 11:10:58.845: INFO: Pod "pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1": Phase="Running", Reason="", readiness=true. Elapsed: 2.022036129s + Feb 25 11:11:00.846: INFO: Pod "pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1": Phase="Running", Reason="", readiness=false. Elapsed: 4.022969986s + Feb 25 11:11:02.854: INFO: Pod "pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.030559788s + STEP: Saw pod success 02/25/23 11:11:02.854 + Feb 25 11:11:02.854: INFO: Pod "pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1" satisfied condition "Succeeded or Failed" + Feb 25 11:11:02.863: INFO: Trying to get logs from node diek3dua9gay-2 pod pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1 container agnhost-container: + STEP: delete the pod 02/25/23 11:11:02.974 + Feb 25 11:11:03.000: INFO: Waiting for pod pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1 to disappear + Feb 25 11:11:03.006: INFO: Pod pod-projected-configmaps-96895ff1-148e-4d1b-97cc-810d74d41ed1 no longer exists + [AfterEach] [sig-storage] Projected configMap + test/e2e/framework/framework.go:187 + Feb 25 11:11:03.007: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-5843" for this suite. 02/25/23 11:11:03.017 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSS ------------------------------ -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - works for multiple CRDs of same group and version but different kinds [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:356 -[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +[sig-apps] Deployment + should run the lifecycle of a Deployment [Conformance] + test/e2e/apps/deployment.go:185 +[BeforeEach] [sig-apps] Deployment test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:44:31.331 -Dec 14 16:44:31.331: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 16:44:31.333 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:44:31.358 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:44:31.362 -[It] works for multiple CRDs of same group and version but different kinds [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:356 -STEP: CRs in the same group and version but different kinds (two CRDs) show up in OpenAPI documentation 12/14/22 16:44:31.365 -Dec 14 16:44:31.367: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:44:34.414: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 11:11:03.04 +Feb 25 11:11:03.040: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename deployment 02/25/23 11:11:03.061 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:11:03.093 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:11:03.1 +[BeforeEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:91 +[It] should run the lifecycle of a Deployment [Conformance] + test/e2e/apps/deployment.go:185 +STEP: creating a Deployment 02/25/23 11:11:03.119 +STEP: waiting for Deployment to be created 02/25/23 11:11:03.138 +STEP: waiting for all Replicas to be Ready 02/25/23 11:11:03.145 +Feb 25 11:11:03.158: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] +Feb 25 11:11:03.158: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] +Feb 25 11:11:03.176: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] +Feb 25 11:11:03.176: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] +Feb 25 11:11:03.239: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] +Feb 25 11:11:03.239: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] +Feb 25 11:11:03.341: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] +Feb 25 11:11:03.341: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] +Feb 25 11:11:05.471: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 and labels map[test-deployment-static:true] +Feb 25 11:11:05.471: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 and labels map[test-deployment-static:true] +Feb 25 11:11:05.508: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 and labels map[test-deployment-static:true] +STEP: patching the Deployment 02/25/23 11:11:05.509 +W0225 11:11:05.529770 13 warnings.go:70] unknown field "spec.template.spec.TerminationGracePeriodSeconds" +Feb 25 11:11:05.535: INFO: observed event type ADDED +STEP: waiting for Replicas to scale 02/25/23 11:11:05.535 +Feb 25 11:11:05.539: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 +Feb 25 11:11:05.539: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 +Feb 25 11:11:05.539: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 +Feb 25 11:11:05.539: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 +Feb 25 11:11:05.540: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 +Feb 25 11:11:05.541: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 +Feb 25 11:11:05.542: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 +Feb 25 11:11:05.542: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 +Feb 25 11:11:05.543: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 +Feb 25 11:11:05.543: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 +Feb 25 11:11:05.543: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 +Feb 25 11:11:05.543: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 +Feb 25 11:11:05.544: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 +Feb 25 11:11:05.544: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 +Feb 25 11:11:05.576: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 +Feb 25 11:11:05.576: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 +Feb 25 11:11:05.670: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 +Feb 25 11:11:05.670: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 +Feb 25 11:11:05.785: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 +Feb 25 11:11:05.785: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 +Feb 25 11:11:05.804: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 +Feb 25 11:11:05.804: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 +Feb 25 11:11:08.519: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 +Feb 25 11:11:08.519: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 +Feb 25 11:11:08.582: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 +STEP: listing Deployments 02/25/23 11:11:08.582 +Feb 25 11:11:08.596: INFO: Found test-deployment with labels: map[test-deployment:patched test-deployment-static:true] +STEP: updating the Deployment 02/25/23 11:11:08.596 +Feb 25 11:11:08.644: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 +STEP: fetching the DeploymentStatus 02/25/23 11:11:08.644 +Feb 25 11:11:08.699: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] +Feb 25 11:11:08.700: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] +Feb 25 11:11:08.735: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] +Feb 25 11:11:08.833: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] +Feb 25 11:11:08.847: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] +Feb 25 11:11:10.541: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] +Feb 25 11:11:10.640: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 3 and labels map[test-deployment:updated test-deployment-static:true] +Feb 25 11:11:10.742: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] +Feb 25 11:11:10.806: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] +Feb 25 11:11:13.713: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 3 and labels map[test-deployment:updated test-deployment-static:true] +STEP: patching the DeploymentStatus 02/25/23 11:11:13.885 +STEP: fetching the DeploymentStatus 02/25/23 11:11:13.925 +Feb 25 11:11:13.972: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 +Feb 25 11:11:13.972: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 +Feb 25 11:11:13.973: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 +Feb 25 11:11:13.973: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 +Feb 25 11:11:13.973: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 +Feb 25 11:11:13.973: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 +Feb 25 11:11:13.973: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 3 +Feb 25 11:11:13.974: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 +Feb 25 11:11:13.974: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 +Feb 25 11:11:13.974: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 3 +STEP: deleting the Deployment 02/25/23 11:11:13.975 +Feb 25 11:11:14.010: INFO: observed event type MODIFIED +Feb 25 11:11:14.010: INFO: observed event type MODIFIED +Feb 25 11:11:14.010: INFO: observed event type MODIFIED +Feb 25 11:11:14.010: INFO: observed event type MODIFIED +Feb 25 11:11:14.011: INFO: observed event type MODIFIED +Feb 25 11:11:14.011: INFO: observed event type MODIFIED +Feb 25 11:11:14.011: INFO: observed event type MODIFIED +Feb 25 11:11:14.011: INFO: observed event type MODIFIED +Feb 25 11:11:14.011: INFO: observed event type MODIFIED +Feb 25 11:11:14.012: INFO: observed event type MODIFIED +Feb 25 11:11:14.012: INFO: observed event type MODIFIED +[AfterEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:84 +Feb 25 11:11:14.038: INFO: Log out all the ReplicaSets if there is no deployment created +Feb 25 11:11:14.050: INFO: ReplicaSet "test-deployment-54cc775c4b": +&ReplicaSet{ObjectMeta:{test-deployment-54cc775c4b deployment-2492 d105b360-2072-4454-91b5-7b22d3df9876 35621 4 2023-02-25 11:11:05 +0000 UTC map[pod-template-hash:54cc775c4b test-deployment-static:true] map[deployment.kubernetes.io/desired-replicas:2 deployment.kubernetes.io/max-replicas:3 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-deployment e50795f8-3e12-4d30-91e5-84e08f6bca8a 0xc006747407 0xc006747408}] [] [{kube-controller-manager Update apps/v1 2023-02-25 11:11:13 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"e50795f8-3e12-4d30-91e5-84e08f6bca8a\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 11:11:13 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{pod-template-hash: 54cc775c4b,test-deployment-static: true,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[pod-template-hash:54cc775c4b test-deployment-static:true] map[] [] [] []} {[] [] [{test-deployment registry.k8s.io/pause:3.8 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc006747490 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:4,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} + +Feb 25 11:11:14.063: INFO: pod: "test-deployment-54cc775c4b-zh5m6": +&Pod{ObjectMeta:{test-deployment-54cc775c4b-zh5m6 test-deployment-54cc775c4b- deployment-2492 166e8720-3b20-44f2-ac18-652f73c61175 35617 0 2023-02-25 11:11:08 +0000 UTC 2023-02-25 11:11:14 +0000 UTC 0xc002ae4b28 map[pod-template-hash:54cc775c4b test-deployment-static:true] map[] [{apps/v1 ReplicaSet test-deployment-54cc775c4b d105b360-2072-4454-91b5-7b22d3df9876 0xc002ae4b57 0xc002ae4b58}] [] [{kube-controller-manager Update v1 2023-02-25 11:11:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"d105b360-2072-4454-91b5-7b22d3df9876\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 11:11:10 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.171\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-flhld,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:test-deployment,Image:registry.k8s.io/pause:3.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-flhld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*1,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:10.233.64.171,StartTime:2023-02-25 11:11:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:test-deployment,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 11:11:10 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/pause:3.8,ImageID:registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d,ContainerID:cri-o://0468c9b67d8e7122643cb06f201d445132b0b58d9eb9247439a5a1f58ddc5bc9,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.171,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + +Feb 25 11:11:14.064: INFO: ReplicaSet "test-deployment-7c7d8d58c8": +&ReplicaSet{ObjectMeta:{test-deployment-7c7d8d58c8 deployment-2492 456b4066-922b-4f7a-bf1c-7c06538fac77 35613 2 2023-02-25 11:11:08 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[deployment.kubernetes.io/desired-replicas:2 deployment.kubernetes.io/max-replicas:3 deployment.kubernetes.io/revision:3] [{apps/v1 Deployment test-deployment e50795f8-3e12-4d30-91e5-84e08f6bca8a 0xc006747517 0xc006747518}] [] [{kube-controller-manager Update apps/v1 2023-02-25 11:11:10 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"e50795f8-3e12-4d30-91e5-84e08f6bca8a\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 11:11:13 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*2,Selector:&v1.LabelSelector{MatchLabels:map[string]string{pod-template-hash: 7c7d8d58c8,test-deployment-static: true,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[] [] [] []} {[] [] [{test-deployment registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0067475a0 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:2,FullyLabeledReplicas:2,ObservedGeneration:2,ReadyReplicas:2,AvailableReplicas:2,Conditions:[]ReplicaSetCondition{},},} + +Feb 25 11:11:14.104: INFO: pod: "test-deployment-7c7d8d58c8-rxwh8": +&Pod{ObjectMeta:{test-deployment-7c7d8d58c8-rxwh8 test-deployment-7c7d8d58c8- deployment-2492 88249978-0da8-48c9-912e-a1fcb924c4c3 35575 0 2023-02-25 11:11:08 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[] [{apps/v1 ReplicaSet test-deployment-7c7d8d58c8 456b4066-922b-4f7a-bf1c-7c06538fac77 0xc002ae5967 0xc002ae5968}] [] [{kube-controller-manager Update v1 2023-02-25 11:11:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"456b4066-922b-4f7a-bf1c-7c06538fac77\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 11:11:10 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.129\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-xpnzg,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:test-deployment,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xpnzg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*1,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:10.233.66.129,StartTime:2023-02-25 11:11:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:test-deployment,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 11:11:10 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://aa8069c391a1b4d9820f9decd3317e2a2601a47fe156b17643f46fb97eb939d7,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.129,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + +Feb 25 11:11:14.104: INFO: pod: "test-deployment-7c7d8d58c8-t4r52": +&Pod{ObjectMeta:{test-deployment-7c7d8d58c8-t4r52 test-deployment-7c7d8d58c8- deployment-2492 3310fc6a-b581-4d27-a3dc-1d4dad0372f9 35612 0 2023-02-25 11:11:10 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[] [{apps/v1 ReplicaSet test-deployment-7c7d8d58c8 456b4066-922b-4f7a-bf1c-7c06538fac77 0xc002ae5d17 0xc002ae5d18}] [] [{kube-controller-manager Update v1 2023-02-25 11:11:10 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"456b4066-922b-4f7a-bf1c-7c06538fac77\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 11:11:13 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.45\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-srxzb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:test-deployment,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-srxzb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*1,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:13 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:13 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:10 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:10.233.65.45,StartTime:2023-02-25 11:11:10 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:test-deployment,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 11:11:12 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://a8f01677774a50d8d2a207cf007066709760c6df5165da09742539ddcc564fea,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.45,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + +Feb 25 11:11:14.104: INFO: ReplicaSet "test-deployment-8594bb6fdd": +&ReplicaSet{ObjectMeta:{test-deployment-8594bb6fdd deployment-2492 067f7061-de9f-4fe7-adc9-9819dd2fed4e 35527 3 2023-02-25 11:11:03 +0000 UTC map[pod-template-hash:8594bb6fdd test-deployment-static:true] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-deployment e50795f8-3e12-4d30-91e5-84e08f6bca8a 0xc006747607 0xc006747608}] [] [{kube-controller-manager Update apps/v1 2023-02-25 11:11:08 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"e50795f8-3e12-4d30-91e5-84e08f6bca8a\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 11:11:08 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{pod-template-hash: 8594bb6fdd,test-deployment-static: true,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[pod-template-hash:8594bb6fdd test-deployment-static:true] map[] [] [] []} {[] [] [{test-deployment registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc006747690 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} + +[AfterEach] [sig-apps] Deployment test/e2e/framework/framework.go:187 -Dec 14 16:44:48.864: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-publish-openapi-7220" for this suite. 12/14/22 16:44:48.878 -{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group and version but different kinds [Conformance]","completed":235,"skipped":4387,"failed":0} +Feb 25 11:11:14.120: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "deployment-2492" for this suite. 02/25/23 11:11:14.167 +{"msg":"PASSED [sig-apps] Deployment should run the lifecycle of a Deployment [Conformance]","completed":231,"skipped":4274,"failed":0} ------------------------------ -• [SLOW TEST] [17.563 seconds] -[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - works for multiple CRDs of same group and version but different kinds [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:356 +• [SLOW TEST] [11.140 seconds] +[sig-apps] Deployment +test/e2e/apps/framework.go:23 + should run the lifecycle of a Deployment [Conformance] + test/e2e/apps/deployment.go:185 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + [BeforeEach] [sig-apps] Deployment test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:44:31.331 - Dec 14 16:44:31.331: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 16:44:31.333 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:44:31.358 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:44:31.362 - [It] works for multiple CRDs of same group and version but different kinds [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:356 - STEP: CRs in the same group and version but different kinds (two CRDs) show up in OpenAPI documentation 12/14/22 16:44:31.365 - Dec 14 16:44:31.367: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:44:34.414: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 11:11:03.04 + Feb 25 11:11:03.040: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename deployment 02/25/23 11:11:03.061 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:11:03.093 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:11:03.1 + [BeforeEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:91 + [It] should run the lifecycle of a Deployment [Conformance] + test/e2e/apps/deployment.go:185 + STEP: creating a Deployment 02/25/23 11:11:03.119 + STEP: waiting for Deployment to be created 02/25/23 11:11:03.138 + STEP: waiting for all Replicas to be Ready 02/25/23 11:11:03.145 + Feb 25 11:11:03.158: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] + Feb 25 11:11:03.158: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] + Feb 25 11:11:03.176: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] + Feb 25 11:11:03.176: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] + Feb 25 11:11:03.239: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] + Feb 25 11:11:03.239: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] + Feb 25 11:11:03.341: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] + Feb 25 11:11:03.341: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 and labels map[test-deployment-static:true] + Feb 25 11:11:05.471: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 and labels map[test-deployment-static:true] + Feb 25 11:11:05.471: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 and labels map[test-deployment-static:true] + Feb 25 11:11:05.508: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 and labels map[test-deployment-static:true] + STEP: patching the Deployment 02/25/23 11:11:05.509 + W0225 11:11:05.529770 13 warnings.go:70] unknown field "spec.template.spec.TerminationGracePeriodSeconds" + Feb 25 11:11:05.535: INFO: observed event type ADDED + STEP: waiting for Replicas to scale 02/25/23 11:11:05.535 + Feb 25 11:11:05.539: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 + Feb 25 11:11:05.539: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 + Feb 25 11:11:05.539: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 + Feb 25 11:11:05.539: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 + Feb 25 11:11:05.540: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 + Feb 25 11:11:05.541: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 + Feb 25 11:11:05.542: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 + Feb 25 11:11:05.542: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 0 + Feb 25 11:11:05.543: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 + Feb 25 11:11:05.543: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 + Feb 25 11:11:05.543: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 + Feb 25 11:11:05.543: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 + Feb 25 11:11:05.544: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 + Feb 25 11:11:05.544: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 + Feb 25 11:11:05.576: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 + Feb 25 11:11:05.576: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 + Feb 25 11:11:05.670: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 + Feb 25 11:11:05.670: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 + Feb 25 11:11:05.785: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 + Feb 25 11:11:05.785: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 + Feb 25 11:11:05.804: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 + Feb 25 11:11:05.804: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 + Feb 25 11:11:08.519: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 + Feb 25 11:11:08.519: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 + Feb 25 11:11:08.582: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 + STEP: listing Deployments 02/25/23 11:11:08.582 + Feb 25 11:11:08.596: INFO: Found test-deployment with labels: map[test-deployment:patched test-deployment-static:true] + STEP: updating the Deployment 02/25/23 11:11:08.596 + Feb 25 11:11:08.644: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 + STEP: fetching the DeploymentStatus 02/25/23 11:11:08.644 + Feb 25 11:11:08.699: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] + Feb 25 11:11:08.700: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] + Feb 25 11:11:08.735: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] + Feb 25 11:11:08.833: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] + Feb 25 11:11:08.847: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 and labels map[test-deployment:updated test-deployment-static:true] + Feb 25 11:11:10.541: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] + Feb 25 11:11:10.640: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 3 and labels map[test-deployment:updated test-deployment-static:true] + Feb 25 11:11:10.742: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] + Feb 25 11:11:10.806: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 and labels map[test-deployment:updated test-deployment-static:true] + Feb 25 11:11:13.713: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 3 and labels map[test-deployment:updated test-deployment-static:true] + STEP: patching the DeploymentStatus 02/25/23 11:11:13.885 + STEP: fetching the DeploymentStatus 02/25/23 11:11:13.925 + Feb 25 11:11:13.972: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 + Feb 25 11:11:13.972: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 + Feb 25 11:11:13.973: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 + Feb 25 11:11:13.973: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 + Feb 25 11:11:13.973: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 1 + Feb 25 11:11:13.973: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 + Feb 25 11:11:13.973: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 3 + Feb 25 11:11:13.974: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 + Feb 25 11:11:13.974: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 2 + Feb 25 11:11:13.974: INFO: observed Deployment test-deployment in namespace deployment-2492 with ReadyReplicas 3 + STEP: deleting the Deployment 02/25/23 11:11:13.975 + Feb 25 11:11:14.010: INFO: observed event type MODIFIED + Feb 25 11:11:14.010: INFO: observed event type MODIFIED + Feb 25 11:11:14.010: INFO: observed event type MODIFIED + Feb 25 11:11:14.010: INFO: observed event type MODIFIED + Feb 25 11:11:14.011: INFO: observed event type MODIFIED + Feb 25 11:11:14.011: INFO: observed event type MODIFIED + Feb 25 11:11:14.011: INFO: observed event type MODIFIED + Feb 25 11:11:14.011: INFO: observed event type MODIFIED + Feb 25 11:11:14.011: INFO: observed event type MODIFIED + Feb 25 11:11:14.012: INFO: observed event type MODIFIED + Feb 25 11:11:14.012: INFO: observed event type MODIFIED + [AfterEach] [sig-apps] Deployment + test/e2e/apps/deployment.go:84 + Feb 25 11:11:14.038: INFO: Log out all the ReplicaSets if there is no deployment created + Feb 25 11:11:14.050: INFO: ReplicaSet "test-deployment-54cc775c4b": + &ReplicaSet{ObjectMeta:{test-deployment-54cc775c4b deployment-2492 d105b360-2072-4454-91b5-7b22d3df9876 35621 4 2023-02-25 11:11:05 +0000 UTC map[pod-template-hash:54cc775c4b test-deployment-static:true] map[deployment.kubernetes.io/desired-replicas:2 deployment.kubernetes.io/max-replicas:3 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-deployment e50795f8-3e12-4d30-91e5-84e08f6bca8a 0xc006747407 0xc006747408}] [] [{kube-controller-manager Update apps/v1 2023-02-25 11:11:13 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"e50795f8-3e12-4d30-91e5-84e08f6bca8a\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 11:11:13 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{pod-template-hash: 54cc775c4b,test-deployment-static: true,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[pod-template-hash:54cc775c4b test-deployment-static:true] map[] [] [] []} {[] [] [{test-deployment registry.k8s.io/pause:3.8 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc006747490 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:4,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} + + Feb 25 11:11:14.063: INFO: pod: "test-deployment-54cc775c4b-zh5m6": + &Pod{ObjectMeta:{test-deployment-54cc775c4b-zh5m6 test-deployment-54cc775c4b- deployment-2492 166e8720-3b20-44f2-ac18-652f73c61175 35617 0 2023-02-25 11:11:08 +0000 UTC 2023-02-25 11:11:14 +0000 UTC 0xc002ae4b28 map[pod-template-hash:54cc775c4b test-deployment-static:true] map[] [{apps/v1 ReplicaSet test-deployment-54cc775c4b d105b360-2072-4454-91b5-7b22d3df9876 0xc002ae4b57 0xc002ae4b58}] [] [{kube-controller-manager Update v1 2023-02-25 11:11:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"d105b360-2072-4454-91b5-7b22d3df9876\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 11:11:10 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.171\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-flhld,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:test-deployment,Image:registry.k8s.io/pause:3.8,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-flhld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*1,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.117,PodIP:10.233.64.171,StartTime:2023-02-25 11:11:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:test-deployment,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 11:11:10 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/pause:3.8,ImageID:registry.k8s.io/pause@sha256:9001185023633d17a2f98ff69b6ff2615b8ea02a825adffa40422f51dfdcde9d,ContainerID:cri-o://0468c9b67d8e7122643cb06f201d445132b0b58d9eb9247439a5a1f58ddc5bc9,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.64.171,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + + Feb 25 11:11:14.064: INFO: ReplicaSet "test-deployment-7c7d8d58c8": + &ReplicaSet{ObjectMeta:{test-deployment-7c7d8d58c8 deployment-2492 456b4066-922b-4f7a-bf1c-7c06538fac77 35613 2 2023-02-25 11:11:08 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[deployment.kubernetes.io/desired-replicas:2 deployment.kubernetes.io/max-replicas:3 deployment.kubernetes.io/revision:3] [{apps/v1 Deployment test-deployment e50795f8-3e12-4d30-91e5-84e08f6bca8a 0xc006747517 0xc006747518}] [] [{kube-controller-manager Update apps/v1 2023-02-25 11:11:10 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"e50795f8-3e12-4d30-91e5-84e08f6bca8a\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 11:11:13 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*2,Selector:&v1.LabelSelector{MatchLabels:map[string]string{pod-template-hash: 7c7d8d58c8,test-deployment-static: true,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[] [] [] []} {[] [] [{test-deployment registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0067475a0 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:2,FullyLabeledReplicas:2,ObservedGeneration:2,ReadyReplicas:2,AvailableReplicas:2,Conditions:[]ReplicaSetCondition{},},} + + Feb 25 11:11:14.104: INFO: pod: "test-deployment-7c7d8d58c8-rxwh8": + &Pod{ObjectMeta:{test-deployment-7c7d8d58c8-rxwh8 test-deployment-7c7d8d58c8- deployment-2492 88249978-0da8-48c9-912e-a1fcb924c4c3 35575 0 2023-02-25 11:11:08 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[] [{apps/v1 ReplicaSet test-deployment-7c7d8d58c8 456b4066-922b-4f7a-bf1c-7c06538fac77 0xc002ae5967 0xc002ae5968}] [] [{kube-controller-manager Update v1 2023-02-25 11:11:08 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"456b4066-922b-4f7a-bf1c-7c06538fac77\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 11:11:10 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.129\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-xpnzg,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:test-deployment,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xpnzg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*1,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:08 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:08 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.44,PodIP:10.233.66.129,StartTime:2023-02-25 11:11:08 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:test-deployment,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 11:11:10 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://aa8069c391a1b4d9820f9decd3317e2a2601a47fe156b17643f46fb97eb939d7,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.129,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + + Feb 25 11:11:14.104: INFO: pod: "test-deployment-7c7d8d58c8-t4r52": + &Pod{ObjectMeta:{test-deployment-7c7d8d58c8-t4r52 test-deployment-7c7d8d58c8- deployment-2492 3310fc6a-b581-4d27-a3dc-1d4dad0372f9 35612 0 2023-02-25 11:11:10 +0000 UTC map[pod-template-hash:7c7d8d58c8 test-deployment-static:true] map[] [{apps/v1 ReplicaSet test-deployment-7c7d8d58c8 456b4066-922b-4f7a-bf1c-7c06538fac77 0xc002ae5d17 0xc002ae5d18}] [] [{kube-controller-manager Update v1 2023-02-25 11:11:10 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"456b4066-922b-4f7a-bf1c-7c06538fac77\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2023-02-25 11:11:13 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.45\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-srxzb,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:test-deployment,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-srxzb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*1,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:diek3dua9gay-1,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:10 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:13 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:13 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2023-02-25 11:11:10 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.250,PodIP:10.233.65.45,StartTime:2023-02-25 11:11:10 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:test-deployment,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2023-02-25 11:11:12 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://a8f01677774a50d8d2a207cf007066709760c6df5165da09742539ddcc564fea,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.65.45,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + + Feb 25 11:11:14.104: INFO: ReplicaSet "test-deployment-8594bb6fdd": + &ReplicaSet{ObjectMeta:{test-deployment-8594bb6fdd deployment-2492 067f7061-de9f-4fe7-adc9-9819dd2fed4e 35527 3 2023-02-25 11:11:03 +0000 UTC map[pod-template-hash:8594bb6fdd test-deployment-static:true] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-deployment e50795f8-3e12-4d30-91e5-84e08f6bca8a 0xc006747607 0xc006747608}] [] [{kube-controller-manager Update apps/v1 2023-02-25 11:11:08 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"e50795f8-3e12-4d30-91e5-84e08f6bca8a\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:pod-template-hash":{},"f:test-deployment-static":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"test-deployment\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2023-02-25 11:11:08 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{pod-template-hash: 8594bb6fdd,test-deployment-static: true,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[pod-template-hash:8594bb6fdd test-deployment-static:true] map[] [] [] []} {[] [] [{test-deployment registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc006747690 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:3,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} + + [AfterEach] [sig-apps] Deployment test/e2e/framework/framework.go:187 - Dec 14 16:44:48.864: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "crd-publish-openapi-7220" for this suite. 12/14/22 16:44:48.878 + Feb 25 11:11:14.120: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "deployment-2492" for this suite. 02/25/23 11:11:14.167 << End Captured GinkgoWriter Output ------------------------------ -SSSSSS +SSSSSSSSSSSS ------------------------------ -[sig-node] Pods - should be submitted and removed [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:225 -[BeforeEach] [sig-node] Pods +[sig-apps] DisruptionController + should observe PodDisruptionBudget status updated [Conformance] + test/e2e/apps/disruption.go:140 +[BeforeEach] [sig-apps] DisruptionController test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:44:48.897 -Dec 14 16:44:48.897: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename pods 12/14/22 16:44:48.899 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:44:48.924 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:44:48.928 -[BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 -[It] should be submitted and removed [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:225 -STEP: creating the pod 12/14/22 16:44:48.931 -STEP: setting up watch 12/14/22 16:44:48.932 -STEP: submitting the pod to kubernetes 12/14/22 16:44:49.038 -STEP: verifying the pod is in kubernetes 12/14/22 16:44:49.058 -STEP: verifying pod creation was observed 12/14/22 16:44:49.063 -Dec 14 16:44:49.063: INFO: Waiting up to 5m0s for pod "pod-submit-remove-4b3a5c8d-a8a8-4086-b5d6-8159a660a349" in namespace "pods-3049" to be "running" -Dec 14 16:44:49.099: INFO: Pod "pod-submit-remove-4b3a5c8d-a8a8-4086-b5d6-8159a660a349": Phase="Pending", Reason="", readiness=false. Elapsed: 35.410427ms -Dec 14 16:44:51.115: INFO: Pod "pod-submit-remove-4b3a5c8d-a8a8-4086-b5d6-8159a660a349": Phase="Running", Reason="", readiness=true. Elapsed: 2.05199155s -Dec 14 16:44:51.116: INFO: Pod "pod-submit-remove-4b3a5c8d-a8a8-4086-b5d6-8159a660a349" satisfied condition "running" -STEP: deleting the pod gracefully 12/14/22 16:44:51.123 -STEP: verifying pod deletion was observed 12/14/22 16:44:51.135 -[AfterEach] [sig-node] Pods +STEP: Creating a kubernetes client 02/25/23 11:11:14.182 +Feb 25 11:11:14.182: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename disruption 02/25/23 11:11:14.193 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:11:14.229 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:11:14.232 +[BeforeEach] [sig-apps] DisruptionController + test/e2e/apps/disruption.go:71 +[It] should observe PodDisruptionBudget status updated [Conformance] + test/e2e/apps/disruption.go:140 +STEP: Waiting for the pdb to be processed 02/25/23 11:11:14.25 +STEP: Waiting for all pods to be running 02/25/23 11:11:16.319 +Feb 25 11:11:16.329: INFO: running pods: 0 < 3 +Feb 25 11:11:18.339: INFO: running pods: 0 < 3 +[AfterEach] [sig-apps] DisruptionController test/e2e/framework/framework.go:187 -Dec 14 16:44:53.680: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-3049" for this suite. 12/14/22 16:44:53.687 -{"msg":"PASSED [sig-node] Pods should be submitted and removed [NodeConformance] [Conformance]","completed":236,"skipped":4393,"failed":0} +Feb 25 11:11:20.347: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "disruption-8596" for this suite. 02/25/23 11:11:20.362 +{"msg":"PASSED [sig-apps] DisruptionController should observe PodDisruptionBudget status updated [Conformance]","completed":232,"skipped":4286,"failed":0} ------------------------------ -• [4.802 seconds] -[sig-node] Pods -test/e2e/common/node/framework.go:23 - should be submitted and removed [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:225 +• [SLOW TEST] [6.197 seconds] +[sig-apps] DisruptionController +test/e2e/apps/framework.go:23 + should observe PodDisruptionBudget status updated [Conformance] + test/e2e/apps/disruption.go:140 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Pods + [BeforeEach] [sig-apps] DisruptionController test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:44:48.897 - Dec 14 16:44:48.897: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename pods 12/14/22 16:44:48.899 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:44:48.924 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:44:48.928 - [BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 - [It] should be submitted and removed [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:225 - STEP: creating the pod 12/14/22 16:44:48.931 - STEP: setting up watch 12/14/22 16:44:48.932 - STEP: submitting the pod to kubernetes 12/14/22 16:44:49.038 - STEP: verifying the pod is in kubernetes 12/14/22 16:44:49.058 - STEP: verifying pod creation was observed 12/14/22 16:44:49.063 - Dec 14 16:44:49.063: INFO: Waiting up to 5m0s for pod "pod-submit-remove-4b3a5c8d-a8a8-4086-b5d6-8159a660a349" in namespace "pods-3049" to be "running" - Dec 14 16:44:49.099: INFO: Pod "pod-submit-remove-4b3a5c8d-a8a8-4086-b5d6-8159a660a349": Phase="Pending", Reason="", readiness=false. Elapsed: 35.410427ms - Dec 14 16:44:51.115: INFO: Pod "pod-submit-remove-4b3a5c8d-a8a8-4086-b5d6-8159a660a349": Phase="Running", Reason="", readiness=true. Elapsed: 2.05199155s - Dec 14 16:44:51.116: INFO: Pod "pod-submit-remove-4b3a5c8d-a8a8-4086-b5d6-8159a660a349" satisfied condition "running" - STEP: deleting the pod gracefully 12/14/22 16:44:51.123 - STEP: verifying pod deletion was observed 12/14/22 16:44:51.135 - [AfterEach] [sig-node] Pods + STEP: Creating a kubernetes client 02/25/23 11:11:14.182 + Feb 25 11:11:14.182: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename disruption 02/25/23 11:11:14.193 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:11:14.229 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:11:14.232 + [BeforeEach] [sig-apps] DisruptionController + test/e2e/apps/disruption.go:71 + [It] should observe PodDisruptionBudget status updated [Conformance] + test/e2e/apps/disruption.go:140 + STEP: Waiting for the pdb to be processed 02/25/23 11:11:14.25 + STEP: Waiting for all pods to be running 02/25/23 11:11:16.319 + Feb 25 11:11:16.329: INFO: running pods: 0 < 3 + Feb 25 11:11:18.339: INFO: running pods: 0 < 3 + [AfterEach] [sig-apps] DisruptionController test/e2e/framework/framework.go:187 - Dec 14 16:44:53.680: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "pods-3049" for this suite. 12/14/22 16:44:53.687 + Feb 25 11:11:20.347: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "disruption-8596" for this suite. 02/25/23 11:11:20.362 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSS +SSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] RuntimeClass - should reject a Pod requesting a non-existent RuntimeClass [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:55 -[BeforeEach] [sig-node] RuntimeClass +[sig-storage] EmptyDir volumes + should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:176 +[BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:44:53.71 -Dec 14 16:44:53.711: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename runtimeclass 12/14/22 16:44:53.715 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:44:53.744 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:44:53.748 -[It] should reject a Pod requesting a non-existent RuntimeClass [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:55 -[AfterEach] [sig-node] RuntimeClass +STEP: Creating a kubernetes client 02/25/23 11:11:20.387 +Feb 25 11:11:20.388: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir 02/25/23 11:11:20.392 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:11:20.435 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:11:20.442 +[It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:176 +STEP: Creating a pod to test emptydir 0666 on node default medium 02/25/23 11:11:20.452 +Feb 25 11:11:20.470: INFO: Waiting up to 5m0s for pod "pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c" in namespace "emptydir-8358" to be "Succeeded or Failed" +Feb 25 11:11:20.479: INFO: Pod "pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c": Phase="Pending", Reason="", readiness=false. Elapsed: 8.57135ms +Feb 25 11:11:22.488: INFO: Pod "pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01736455s +Feb 25 11:11:24.489: INFO: Pod "pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.018364922s +Feb 25 11:11:26.488: INFO: Pod "pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.018262715s +STEP: Saw pod success 02/25/23 11:11:26.488 +Feb 25 11:11:26.489: INFO: Pod "pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c" satisfied condition "Succeeded or Failed" +Feb 25 11:11:26.497: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c container test-container: +STEP: delete the pod 02/25/23 11:11:26.52 +Feb 25 11:11:26.544: INFO: Waiting for pod pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c to disappear +Feb 25 11:11:26.562: INFO: Pod pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c no longer exists +[AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 -Dec 14 16:44:53.762: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "runtimeclass-3482" for this suite. 12/14/22 16:44:53.772 -{"msg":"PASSED [sig-node] RuntimeClass should reject a Pod requesting a non-existent RuntimeClass [NodeConformance] [Conformance]","completed":237,"skipped":4401,"failed":0} +Feb 25 11:11:26.562: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-8358" for this suite. 02/25/23 11:11:26.586 +{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance]","completed":233,"skipped":4306,"failed":0} ------------------------------ -• [0.087 seconds] -[sig-node] RuntimeClass -test/e2e/common/node/framework.go:23 - should reject a Pod requesting a non-existent RuntimeClass [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:55 +• [SLOW TEST] [6.221 seconds] +[sig-storage] EmptyDir volumes +test/e2e/common/storage/framework.go:23 + should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:176 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] RuntimeClass + [BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:44:53.71 - Dec 14 16:44:53.711: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename runtimeclass 12/14/22 16:44:53.715 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:44:53.744 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:44:53.748 - [It] should reject a Pod requesting a non-existent RuntimeClass [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:55 - [AfterEach] [sig-node] RuntimeClass + STEP: Creating a kubernetes client 02/25/23 11:11:20.387 + Feb 25 11:11:20.388: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir 02/25/23 11:11:20.392 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:11:20.435 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:11:20.442 + [It] should support (root,0666,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:176 + STEP: Creating a pod to test emptydir 0666 on node default medium 02/25/23 11:11:20.452 + Feb 25 11:11:20.470: INFO: Waiting up to 5m0s for pod "pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c" in namespace "emptydir-8358" to be "Succeeded or Failed" + Feb 25 11:11:20.479: INFO: Pod "pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c": Phase="Pending", Reason="", readiness=false. Elapsed: 8.57135ms + Feb 25 11:11:22.488: INFO: Pod "pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01736455s + Feb 25 11:11:24.489: INFO: Pod "pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.018364922s + Feb 25 11:11:26.488: INFO: Pod "pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.018262715s + STEP: Saw pod success 02/25/23 11:11:26.488 + Feb 25 11:11:26.489: INFO: Pod "pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c" satisfied condition "Succeeded or Failed" + Feb 25 11:11:26.497: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c container test-container: + STEP: delete the pod 02/25/23 11:11:26.52 + Feb 25 11:11:26.544: INFO: Waiting for pod pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c to disappear + Feb 25 11:11:26.562: INFO: Pod pod-519227f7-5a9f-4c9f-9ed0-99e07ffacb9c no longer exists + [AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 - Dec 14 16:44:53.762: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "runtimeclass-3482" for this suite. 12/14/22 16:44:53.772 + Feb 25 11:11:26.562: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-8358" for this suite. 02/25/23 11:11:26.586 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] Watchers - should receive events on concurrent watches in same order [Conformance] - test/e2e/apimachinery/watch.go:334 -[BeforeEach] [sig-api-machinery] Watchers +[sig-network] DNS + should provide DNS for ExternalName services [Conformance] + test/e2e/network/dns.go:333 +[BeforeEach] [sig-network] DNS test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:44:53.808 -Dec 14 16:44:53.809: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename watch 12/14/22 16:44:53.81 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:44:53.838 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:44:53.842 -[It] should receive events on concurrent watches in same order [Conformance] - test/e2e/apimachinery/watch.go:334 -STEP: getting a starting resourceVersion 12/14/22 16:44:53.845 -STEP: starting a background goroutine to produce watch events 12/14/22 16:44:53.85 -STEP: creating watches starting from each resource version of the events produced and verifying they all receive resource versions in the same order 12/14/22 16:44:53.851 -[AfterEach] [sig-api-machinery] Watchers +STEP: Creating a kubernetes client 02/25/23 11:11:26.621 +Feb 25 11:11:26.622: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename dns 02/25/23 11:11:26.628 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:11:26.677 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:11:26.681 +[It] should provide DNS for ExternalName services [Conformance] + test/e2e/network/dns.go:333 +STEP: Creating a test externalName service 02/25/23 11:11:26.689 +STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2045.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local; sleep 1; done + 02/25/23 11:11:26.711 +STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2045.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local; sleep 1; done + 02/25/23 11:11:26.711 +STEP: creating a pod to probe DNS 02/25/23 11:11:26.712 +STEP: submitting the pod to kubernetes 02/25/23 11:11:26.712 +Feb 25 11:11:26.748: INFO: Waiting up to 15m0s for pod "dns-test-6241e81c-035b-422e-9b3b-1fcb3c8c527f" in namespace "dns-2045" to be "running" +Feb 25 11:11:26.764: INFO: Pod "dns-test-6241e81c-035b-422e-9b3b-1fcb3c8c527f": Phase="Pending", Reason="", readiness=false. Elapsed: 15.101419ms +Feb 25 11:11:28.775: INFO: Pod "dns-test-6241e81c-035b-422e-9b3b-1fcb3c8c527f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026140737s +Feb 25 11:11:30.771: INFO: Pod "dns-test-6241e81c-035b-422e-9b3b-1fcb3c8c527f": Phase="Running", Reason="", readiness=true. Elapsed: 4.022360497s +Feb 25 11:11:30.771: INFO: Pod "dns-test-6241e81c-035b-422e-9b3b-1fcb3c8c527f" satisfied condition "running" +STEP: retrieving the pod 02/25/23 11:11:30.771 +STEP: looking for the results for each expected name from probers 02/25/23 11:11:30.783 +Feb 25 11:11:30.821: INFO: DNS probes using dns-test-6241e81c-035b-422e-9b3b-1fcb3c8c527f succeeded + +STEP: deleting the pod 02/25/23 11:11:30.821 +STEP: changing the externalName to bar.example.com 02/25/23 11:11:30.851 +STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2045.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local; sleep 1; done + 02/25/23 11:11:30.928 +STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2045.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local; sleep 1; done + 02/25/23 11:11:30.929 +STEP: creating a second pod to probe DNS 02/25/23 11:11:30.93 +STEP: submitting the pod to kubernetes 02/25/23 11:11:30.93 +Feb 25 11:11:30.955: INFO: Waiting up to 15m0s for pod "dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d" in namespace "dns-2045" to be "running" +Feb 25 11:11:30.971: INFO: Pod "dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d": Phase="Pending", Reason="", readiness=false. Elapsed: 15.839311ms +Feb 25 11:11:32.979: INFO: Pod "dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02376326s +Feb 25 11:11:34.982: INFO: Pod "dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d": Phase="Running", Reason="", readiness=true. Elapsed: 4.026277029s +Feb 25 11:11:34.982: INFO: Pod "dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d" satisfied condition "running" +STEP: retrieving the pod 02/25/23 11:11:34.982 +STEP: looking for the results for each expected name from probers 02/25/23 11:11:34.993 +Feb 25 11:11:35.012: INFO: File wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. +' instead of 'bar.example.com.' +Feb 25 11:11:35.024: INFO: File jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. +' instead of 'bar.example.com.' +Feb 25 11:11:35.024: INFO: Lookups using dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d failed for: [wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local] + +Feb 25 11:11:40.036: INFO: File wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. +' instead of 'bar.example.com.' +Feb 25 11:11:40.044: INFO: File jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. +' instead of 'bar.example.com.' +Feb 25 11:11:40.045: INFO: Lookups using dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d failed for: [wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local] + +Feb 25 11:11:45.035: INFO: File wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. +' instead of 'bar.example.com.' +Feb 25 11:11:45.047: INFO: File jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. +' instead of 'bar.example.com.' +Feb 25 11:11:45.047: INFO: Lookups using dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d failed for: [wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local] + +Feb 25 11:11:50.034: INFO: File wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. +' instead of 'bar.example.com.' +Feb 25 11:11:50.041: INFO: File jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. +' instead of 'bar.example.com.' +Feb 25 11:11:50.041: INFO: Lookups using dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d failed for: [wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local] + +Feb 25 11:11:55.036: INFO: File wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. +' instead of 'bar.example.com.' +Feb 25 11:11:55.045: INFO: File jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. +' instead of 'bar.example.com.' +Feb 25 11:11:55.045: INFO: Lookups using dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d failed for: [wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local] + +Feb 25 11:12:00.043: INFO: File jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains '' instead of 'bar.example.com.' +Feb 25 11:12:00.044: INFO: Lookups using dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d failed for: [jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local] + +Feb 25 11:12:05.075: INFO: DNS probes using dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d succeeded + +STEP: deleting the pod 02/25/23 11:12:05.076 +STEP: changing the service to type=ClusterIP 02/25/23 11:12:05.205 +STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2045.svc.cluster.local A > /results/wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local; sleep 1; done + 02/25/23 11:12:05.295 +STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2045.svc.cluster.local A > /results/jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local; sleep 1; done + 02/25/23 11:12:05.295 +STEP: creating a third pod to probe DNS 02/25/23 11:12:05.296 +STEP: submitting the pod to kubernetes 02/25/23 11:12:05.331 +Feb 25 11:12:05.360: INFO: Waiting up to 15m0s for pod "dns-test-0b97ee3c-2730-4b77-8467-0b51a7442258" in namespace "dns-2045" to be "running" +Feb 25 11:12:05.382: INFO: Pod "dns-test-0b97ee3c-2730-4b77-8467-0b51a7442258": Phase="Pending", Reason="", readiness=false. Elapsed: 22.41684ms +Feb 25 11:12:07.391: INFO: Pod "dns-test-0b97ee3c-2730-4b77-8467-0b51a7442258": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031565966s +Feb 25 11:12:09.394: INFO: Pod "dns-test-0b97ee3c-2730-4b77-8467-0b51a7442258": Phase="Running", Reason="", readiness=true. Elapsed: 4.033946754s +Feb 25 11:12:09.394: INFO: Pod "dns-test-0b97ee3c-2730-4b77-8467-0b51a7442258" satisfied condition "running" +STEP: retrieving the pod 02/25/23 11:12:09.394 +STEP: looking for the results for each expected name from probers 02/25/23 11:12:09.4 +Feb 25 11:12:09.424: INFO: DNS probes using dns-test-0b97ee3c-2730-4b77-8467-0b51a7442258 succeeded + +STEP: deleting the pod 02/25/23 11:12:09.424 +STEP: deleting the test externalName service 02/25/23 11:12:09.46 +[AfterEach] [sig-network] DNS test/e2e/framework/framework.go:187 -Dec 14 16:44:56.620: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "watch-4004" for this suite. 12/14/22 16:44:56.67 -{"msg":"PASSED [sig-api-machinery] Watchers should receive events on concurrent watches in same order [Conformance]","completed":238,"skipped":4429,"failed":0} +Feb 25 11:12:09.524: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "dns-2045" for this suite. 02/25/23 11:12:09.54 +{"msg":"PASSED [sig-network] DNS should provide DNS for ExternalName services [Conformance]","completed":234,"skipped":4323,"failed":0} ------------------------------ -• [2.917 seconds] -[sig-api-machinery] Watchers -test/e2e/apimachinery/framework.go:23 - should receive events on concurrent watches in same order [Conformance] - test/e2e/apimachinery/watch.go:334 +• [SLOW TEST] [42.943 seconds] +[sig-network] DNS +test/e2e/network/common/framework.go:23 + should provide DNS for ExternalName services [Conformance] + test/e2e/network/dns.go:333 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Watchers + [BeforeEach] [sig-network] DNS test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:44:53.808 - Dec 14 16:44:53.809: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename watch 12/14/22 16:44:53.81 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:44:53.838 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:44:53.842 - [It] should receive events on concurrent watches in same order [Conformance] - test/e2e/apimachinery/watch.go:334 - STEP: getting a starting resourceVersion 12/14/22 16:44:53.845 - STEP: starting a background goroutine to produce watch events 12/14/22 16:44:53.85 - STEP: creating watches starting from each resource version of the events produced and verifying they all receive resource versions in the same order 12/14/22 16:44:53.851 - [AfterEach] [sig-api-machinery] Watchers + STEP: Creating a kubernetes client 02/25/23 11:11:26.621 + Feb 25 11:11:26.622: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename dns 02/25/23 11:11:26.628 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:11:26.677 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:11:26.681 + [It] should provide DNS for ExternalName services [Conformance] + test/e2e/network/dns.go:333 + STEP: Creating a test externalName service 02/25/23 11:11:26.689 + STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2045.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local; sleep 1; done + 02/25/23 11:11:26.711 + STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2045.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local; sleep 1; done + 02/25/23 11:11:26.711 + STEP: creating a pod to probe DNS 02/25/23 11:11:26.712 + STEP: submitting the pod to kubernetes 02/25/23 11:11:26.712 + Feb 25 11:11:26.748: INFO: Waiting up to 15m0s for pod "dns-test-6241e81c-035b-422e-9b3b-1fcb3c8c527f" in namespace "dns-2045" to be "running" + Feb 25 11:11:26.764: INFO: Pod "dns-test-6241e81c-035b-422e-9b3b-1fcb3c8c527f": Phase="Pending", Reason="", readiness=false. Elapsed: 15.101419ms + Feb 25 11:11:28.775: INFO: Pod "dns-test-6241e81c-035b-422e-9b3b-1fcb3c8c527f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026140737s + Feb 25 11:11:30.771: INFO: Pod "dns-test-6241e81c-035b-422e-9b3b-1fcb3c8c527f": Phase="Running", Reason="", readiness=true. Elapsed: 4.022360497s + Feb 25 11:11:30.771: INFO: Pod "dns-test-6241e81c-035b-422e-9b3b-1fcb3c8c527f" satisfied condition "running" + STEP: retrieving the pod 02/25/23 11:11:30.771 + STEP: looking for the results for each expected name from probers 02/25/23 11:11:30.783 + Feb 25 11:11:30.821: INFO: DNS probes using dns-test-6241e81c-035b-422e-9b3b-1fcb3c8c527f succeeded + + STEP: deleting the pod 02/25/23 11:11:30.821 + STEP: changing the externalName to bar.example.com 02/25/23 11:11:30.851 + STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2045.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local; sleep 1; done + 02/25/23 11:11:30.928 + STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2045.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local; sleep 1; done + 02/25/23 11:11:30.929 + STEP: creating a second pod to probe DNS 02/25/23 11:11:30.93 + STEP: submitting the pod to kubernetes 02/25/23 11:11:30.93 + Feb 25 11:11:30.955: INFO: Waiting up to 15m0s for pod "dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d" in namespace "dns-2045" to be "running" + Feb 25 11:11:30.971: INFO: Pod "dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d": Phase="Pending", Reason="", readiness=false. Elapsed: 15.839311ms + Feb 25 11:11:32.979: INFO: Pod "dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02376326s + Feb 25 11:11:34.982: INFO: Pod "dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d": Phase="Running", Reason="", readiness=true. Elapsed: 4.026277029s + Feb 25 11:11:34.982: INFO: Pod "dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d" satisfied condition "running" + STEP: retrieving the pod 02/25/23 11:11:34.982 + STEP: looking for the results for each expected name from probers 02/25/23 11:11:34.993 + Feb 25 11:11:35.012: INFO: File wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. + ' instead of 'bar.example.com.' + Feb 25 11:11:35.024: INFO: File jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. + ' instead of 'bar.example.com.' + Feb 25 11:11:35.024: INFO: Lookups using dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d failed for: [wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local] + + Feb 25 11:11:40.036: INFO: File wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. + ' instead of 'bar.example.com.' + Feb 25 11:11:40.044: INFO: File jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. + ' instead of 'bar.example.com.' + Feb 25 11:11:40.045: INFO: Lookups using dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d failed for: [wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local] + + Feb 25 11:11:45.035: INFO: File wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. + ' instead of 'bar.example.com.' + Feb 25 11:11:45.047: INFO: File jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. + ' instead of 'bar.example.com.' + Feb 25 11:11:45.047: INFO: Lookups using dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d failed for: [wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local] + + Feb 25 11:11:50.034: INFO: File wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. + ' instead of 'bar.example.com.' + Feb 25 11:11:50.041: INFO: File jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. + ' instead of 'bar.example.com.' + Feb 25 11:11:50.041: INFO: Lookups using dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d failed for: [wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local] + + Feb 25 11:11:55.036: INFO: File wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. + ' instead of 'bar.example.com.' + Feb 25 11:11:55.045: INFO: File jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains 'foo.example.com. + ' instead of 'bar.example.com.' + Feb 25 11:11:55.045: INFO: Lookups using dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d failed for: [wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local] + + Feb 25 11:12:00.043: INFO: File jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local from pod dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d contains '' instead of 'bar.example.com.' + Feb 25 11:12:00.044: INFO: Lookups using dns-2045/dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d failed for: [jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local] + + Feb 25 11:12:05.075: INFO: DNS probes using dns-test-af881579-f1d3-4f9c-b983-39cf9709c07d succeeded + + STEP: deleting the pod 02/25/23 11:12:05.076 + STEP: changing the service to type=ClusterIP 02/25/23 11:12:05.205 + STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2045.svc.cluster.local A > /results/wheezy_udp@dns-test-service-3.dns-2045.svc.cluster.local; sleep 1; done + 02/25/23 11:12:05.295 + STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-2045.svc.cluster.local A > /results/jessie_udp@dns-test-service-3.dns-2045.svc.cluster.local; sleep 1; done + 02/25/23 11:12:05.295 + STEP: creating a third pod to probe DNS 02/25/23 11:12:05.296 + STEP: submitting the pod to kubernetes 02/25/23 11:12:05.331 + Feb 25 11:12:05.360: INFO: Waiting up to 15m0s for pod "dns-test-0b97ee3c-2730-4b77-8467-0b51a7442258" in namespace "dns-2045" to be "running" + Feb 25 11:12:05.382: INFO: Pod "dns-test-0b97ee3c-2730-4b77-8467-0b51a7442258": Phase="Pending", Reason="", readiness=false. Elapsed: 22.41684ms + Feb 25 11:12:07.391: INFO: Pod "dns-test-0b97ee3c-2730-4b77-8467-0b51a7442258": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031565966s + Feb 25 11:12:09.394: INFO: Pod "dns-test-0b97ee3c-2730-4b77-8467-0b51a7442258": Phase="Running", Reason="", readiness=true. Elapsed: 4.033946754s + Feb 25 11:12:09.394: INFO: Pod "dns-test-0b97ee3c-2730-4b77-8467-0b51a7442258" satisfied condition "running" + STEP: retrieving the pod 02/25/23 11:12:09.394 + STEP: looking for the results for each expected name from probers 02/25/23 11:12:09.4 + Feb 25 11:12:09.424: INFO: DNS probes using dns-test-0b97ee3c-2730-4b77-8467-0b51a7442258 succeeded + + STEP: deleting the pod 02/25/23 11:12:09.424 + STEP: deleting the test externalName service 02/25/23 11:12:09.46 + [AfterEach] [sig-network] DNS test/e2e/framework/framework.go:187 - Dec 14 16:44:56.620: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "watch-4004" for this suite. 12/14/22 16:44:56.67 + Feb 25 11:12:09.524: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "dns-2045" for this suite. 02/25/23 11:12:09.54 << End Captured GinkgoWriter Output ------------------------------ SSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] RuntimeClass - should reject a Pod requesting a deleted RuntimeClass [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:156 -[BeforeEach] [sig-node] RuntimeClass +[sig-storage] Projected configMap + should be consumable from pods in volume as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:73 +[BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:44:56.733 -Dec 14 16:44:56.733: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename runtimeclass 12/14/22 16:44:56.735 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:44:56.85 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:44:56.854 -[It] should reject a Pod requesting a deleted RuntimeClass [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:156 -STEP: Deleting RuntimeClass runtimeclass-8597-delete-me 12/14/22 16:44:56.863 -STEP: Waiting for the RuntimeClass to disappear 12/14/22 16:44:56.871 -[AfterEach] [sig-node] RuntimeClass +STEP: Creating a kubernetes client 02/25/23 11:12:09.567 +Feb 25 11:12:09.567: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 11:12:09.57 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:12:09.607 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:12:09.618 +[It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:73 +STEP: Creating configMap with name projected-configmap-test-volume-63ece52b-37f3-4d2a-9612-cf82073757c6 02/25/23 11:12:09.631 +STEP: Creating a pod to test consume configMaps 02/25/23 11:12:09.647 +Feb 25 11:12:09.673: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4" in namespace "projected-2768" to be "Succeeded or Failed" +Feb 25 11:12:09.683: INFO: Pod "pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4": Phase="Pending", Reason="", readiness=false. Elapsed: 9.797811ms +Feb 25 11:12:11.690: INFO: Pod "pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017226826s +Feb 25 11:12:13.690: INFO: Pod "pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4": Phase="Pending", Reason="", readiness=false. Elapsed: 4.016886093s +Feb 25 11:12:15.691: INFO: Pod "pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.0180965s +STEP: Saw pod success 02/25/23 11:12:15.692 +Feb 25 11:12:15.692: INFO: Pod "pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4" satisfied condition "Succeeded or Failed" +Feb 25 11:12:15.698: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4 container agnhost-container: +STEP: delete the pod 02/25/23 11:12:16.08 +Feb 25 11:12:16.100: INFO: Waiting for pod pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4 to disappear +Feb 25 11:12:16.107: INFO: Pod pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4 no longer exists +[AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 -Dec 14 16:44:56.883: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "runtimeclass-8597" for this suite. 12/14/22 16:44:56.889 -{"msg":"PASSED [sig-node] RuntimeClass should reject a Pod requesting a deleted RuntimeClass [NodeConformance] [Conformance]","completed":239,"skipped":4446,"failed":0} +Feb 25 11:12:16.107: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-2768" for this suite. 02/25/23 11:12:16.116 +{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]","completed":235,"skipped":4340,"failed":0} ------------------------------ -• [0.164 seconds] -[sig-node] RuntimeClass -test/e2e/common/node/framework.go:23 - should reject a Pod requesting a deleted RuntimeClass [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:156 +• [SLOW TEST] [6.561 seconds] +[sig-storage] Projected configMap +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:73 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] RuntimeClass + [BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:44:56.733 - Dec 14 16:44:56.733: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename runtimeclass 12/14/22 16:44:56.735 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:44:56.85 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:44:56.854 - [It] should reject a Pod requesting a deleted RuntimeClass [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:156 - STEP: Deleting RuntimeClass runtimeclass-8597-delete-me 12/14/22 16:44:56.863 - STEP: Waiting for the RuntimeClass to disappear 12/14/22 16:44:56.871 - [AfterEach] [sig-node] RuntimeClass + STEP: Creating a kubernetes client 02/25/23 11:12:09.567 + Feb 25 11:12:09.567: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 11:12:09.57 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:12:09.607 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:12:09.618 + [It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:73 + STEP: Creating configMap with name projected-configmap-test-volume-63ece52b-37f3-4d2a-9612-cf82073757c6 02/25/23 11:12:09.631 + STEP: Creating a pod to test consume configMaps 02/25/23 11:12:09.647 + Feb 25 11:12:09.673: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4" in namespace "projected-2768" to be "Succeeded or Failed" + Feb 25 11:12:09.683: INFO: Pod "pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4": Phase="Pending", Reason="", readiness=false. Elapsed: 9.797811ms + Feb 25 11:12:11.690: INFO: Pod "pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017226826s + Feb 25 11:12:13.690: INFO: Pod "pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4": Phase="Pending", Reason="", readiness=false. Elapsed: 4.016886093s + Feb 25 11:12:15.691: INFO: Pod "pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.0180965s + STEP: Saw pod success 02/25/23 11:12:15.692 + Feb 25 11:12:15.692: INFO: Pod "pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4" satisfied condition "Succeeded or Failed" + Feb 25 11:12:15.698: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4 container agnhost-container: + STEP: delete the pod 02/25/23 11:12:16.08 + Feb 25 11:12:16.100: INFO: Waiting for pod pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4 to disappear + Feb 25 11:12:16.107: INFO: Pod pod-projected-configmaps-6aaacbbe-290f-49b2-b594-ae311a1d84a4 no longer exists + [AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 - Dec 14 16:44:56.883: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "runtimeclass-8597" for this suite. 12/14/22 16:44:56.889 + Feb 25 11:12:16.107: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-2768" for this suite. 02/25/23 11:12:16.116 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Variable Expansion - should allow composing env vars into new env vars [NodeConformance] [Conformance] - test/e2e/common/node/expansion.go:43 -[BeforeEach] [sig-node] Variable Expansion +[sig-storage] Secrets + should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:88 +[BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:44:56.899 -Dec 14 16:44:56.899: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename var-expansion 12/14/22 16:44:56.903 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:44:57.029 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:44:57.036 -[It] should allow composing env vars into new env vars [NodeConformance] [Conformance] - test/e2e/common/node/expansion.go:43 -STEP: Creating a pod to test env composition 12/14/22 16:44:57.039 -Dec 14 16:44:57.062: INFO: Waiting up to 5m0s for pod "var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe" in namespace "var-expansion-2651" to be "Succeeded or Failed" -Dec 14 16:44:57.066: INFO: Pod "var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe": Phase="Pending", Reason="", readiness=false. Elapsed: 3.213977ms -Dec 14 16:44:59.071: INFO: Pod "var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008395482s -Dec 14 16:45:01.119: INFO: Pod "var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.056920305s -STEP: Saw pod success 12/14/22 16:45:01.12 -Dec 14 16:45:01.120: INFO: Pod "var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe" satisfied condition "Succeeded or Failed" -Dec 14 16:45:01.124: INFO: Trying to get logs from node kiedai4iakae-3 pod var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe container dapi-container: -STEP: delete the pod 12/14/22 16:45:01.15 -Dec 14 16:45:01.166: INFO: Waiting for pod var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe to disappear -Dec 14 16:45:01.171: INFO: Pod var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe no longer exists -[AfterEach] [sig-node] Variable Expansion +STEP: Creating a kubernetes client 02/25/23 11:12:16.136 +Feb 25 11:12:16.136: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename secrets 02/25/23 11:12:16.137 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:12:16.168 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:12:16.175 +[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:88 +STEP: Creating secret with name secret-test-map-492964f2-4d84-4a10-9df9-2efb830017ce 02/25/23 11:12:16.18 +STEP: Creating a pod to test consume secrets 02/25/23 11:12:16.191 +Feb 25 11:12:16.209: INFO: Waiting up to 5m0s for pod "pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23" in namespace "secrets-4872" to be "Succeeded or Failed" +Feb 25 11:12:16.216: INFO: Pod "pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23": Phase="Pending", Reason="", readiness=false. Elapsed: 6.708956ms +Feb 25 11:12:18.224: INFO: Pod "pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014605554s +Feb 25 11:12:20.222: INFO: Pod "pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23": Phase="Pending", Reason="", readiness=false. Elapsed: 4.013477932s +Feb 25 11:12:22.223: INFO: Pod "pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.014326285s +STEP: Saw pod success 02/25/23 11:12:22.224 +Feb 25 11:12:22.224: INFO: Pod "pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23" satisfied condition "Succeeded or Failed" +Feb 25 11:12:22.229: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23 container secret-volume-test: +STEP: delete the pod 02/25/23 11:12:22.241 +Feb 25 11:12:22.267: INFO: Waiting for pod pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23 to disappear +Feb 25 11:12:22.272: INFO: Pod pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23 no longer exists +[AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 -Dec 14 16:45:01.172: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "var-expansion-2651" for this suite. 12/14/22 16:45:01.178 -{"msg":"PASSED [sig-node] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance]","completed":240,"skipped":4456,"failed":0} +Feb 25 11:12:22.273: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-4872" for this suite. 02/25/23 11:12:22.283 +{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","completed":236,"skipped":4371,"failed":0} ------------------------------ -• [4.287 seconds] -[sig-node] Variable Expansion -test/e2e/common/node/framework.go:23 - should allow composing env vars into new env vars [NodeConformance] [Conformance] - test/e2e/common/node/expansion.go:43 +• [SLOW TEST] [6.160 seconds] +[sig-storage] Secrets +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:88 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Variable Expansion + [BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:44:56.899 - Dec 14 16:44:56.899: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename var-expansion 12/14/22 16:44:56.903 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:44:57.029 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:44:57.036 - [It] should allow composing env vars into new env vars [NodeConformance] [Conformance] - test/e2e/common/node/expansion.go:43 - STEP: Creating a pod to test env composition 12/14/22 16:44:57.039 - Dec 14 16:44:57.062: INFO: Waiting up to 5m0s for pod "var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe" in namespace "var-expansion-2651" to be "Succeeded or Failed" - Dec 14 16:44:57.066: INFO: Pod "var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe": Phase="Pending", Reason="", readiness=false. Elapsed: 3.213977ms - Dec 14 16:44:59.071: INFO: Pod "var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008395482s - Dec 14 16:45:01.119: INFO: Pod "var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.056920305s - STEP: Saw pod success 12/14/22 16:45:01.12 - Dec 14 16:45:01.120: INFO: Pod "var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe" satisfied condition "Succeeded or Failed" - Dec 14 16:45:01.124: INFO: Trying to get logs from node kiedai4iakae-3 pod var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe container dapi-container: - STEP: delete the pod 12/14/22 16:45:01.15 - Dec 14 16:45:01.166: INFO: Waiting for pod var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe to disappear - Dec 14 16:45:01.171: INFO: Pod var-expansion-19f78a18-59e7-4366-9533-69c8ef9865fe no longer exists - [AfterEach] [sig-node] Variable Expansion + STEP: Creating a kubernetes client 02/25/23 11:12:16.136 + Feb 25 11:12:16.136: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename secrets 02/25/23 11:12:16.137 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:12:16.168 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:12:16.175 + [It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:88 + STEP: Creating secret with name secret-test-map-492964f2-4d84-4a10-9df9-2efb830017ce 02/25/23 11:12:16.18 + STEP: Creating a pod to test consume secrets 02/25/23 11:12:16.191 + Feb 25 11:12:16.209: INFO: Waiting up to 5m0s for pod "pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23" in namespace "secrets-4872" to be "Succeeded or Failed" + Feb 25 11:12:16.216: INFO: Pod "pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23": Phase="Pending", Reason="", readiness=false. Elapsed: 6.708956ms + Feb 25 11:12:18.224: INFO: Pod "pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014605554s + Feb 25 11:12:20.222: INFO: Pod "pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23": Phase="Pending", Reason="", readiness=false. Elapsed: 4.013477932s + Feb 25 11:12:22.223: INFO: Pod "pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.014326285s + STEP: Saw pod success 02/25/23 11:12:22.224 + Feb 25 11:12:22.224: INFO: Pod "pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23" satisfied condition "Succeeded or Failed" + Feb 25 11:12:22.229: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23 container secret-volume-test: + STEP: delete the pod 02/25/23 11:12:22.241 + Feb 25 11:12:22.267: INFO: Waiting for pod pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23 to disappear + Feb 25 11:12:22.272: INFO: Pod pod-secrets-343af9fe-56da-408a-8aef-da61b1435d23 no longer exists + [AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 - Dec 14 16:45:01.172: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "var-expansion-2651" for this suite. 12/14/22 16:45:01.178 + Feb 25 11:12:22.273: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "secrets-4872" for this suite. 02/25/23 11:12:22.283 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSS +SSSSSSSSSSS ------------------------------ -[sig-api-machinery] Namespaces [Serial] - should ensure that all pods are removed when a namespace is deleted [Conformance] - test/e2e/apimachinery/namespace.go:242 -[BeforeEach] [sig-api-machinery] Namespaces [Serial] +[sig-node] Ephemeral Containers [NodeConformance] + will start an ephemeral container in an existing pod [Conformance] + test/e2e/common/node/ephemeral_containers.go:45 +[BeforeEach] [sig-node] Ephemeral Containers [NodeConformance] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:45:01.188 -Dec 14 16:45:01.188: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename namespaces 12/14/22 16:45:01.192 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:45:01.22 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:45:01.224 -[It] should ensure that all pods are removed when a namespace is deleted [Conformance] - test/e2e/apimachinery/namespace.go:242 -STEP: Creating a test namespace 12/14/22 16:45:01.228 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:45:01.255 -STEP: Creating a pod in the namespace 12/14/22 16:45:01.259 -STEP: Waiting for the pod to have running status 12/14/22 16:45:01.271 -Dec 14 16:45:01.271: INFO: Waiting up to 5m0s for pod "test-pod" in namespace "nsdeletetest-3234" to be "running" -Dec 14 16:45:01.275: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 3.854922ms -Dec 14 16:45:03.283: INFO: Pod "test-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.011571922s -Dec 14 16:45:03.283: INFO: Pod "test-pod" satisfied condition "running" -STEP: Deleting the namespace 12/14/22 16:45:03.283 -STEP: Waiting for the namespace to be removed. 12/14/22 16:45:03.295 -STEP: Recreating the namespace 12/14/22 16:45:14.302 -STEP: Verifying there are no pods in the namespace 12/14/22 16:45:14.329 -[AfterEach] [sig-api-machinery] Namespaces [Serial] +STEP: Creating a kubernetes client 02/25/23 11:12:22.297 +Feb 25 11:12:22.297: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename ephemeral-containers-test 02/25/23 11:12:22.301 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:12:22.322 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:12:22.328 +[BeforeEach] [sig-node] Ephemeral Containers [NodeConformance] + test/e2e/common/node/ephemeral_containers.go:38 +[It] will start an ephemeral container in an existing pod [Conformance] + test/e2e/common/node/ephemeral_containers.go:45 +STEP: creating a target pod 02/25/23 11:12:22.332 +Feb 25 11:12:22.349: INFO: Waiting up to 5m0s for pod "ephemeral-containers-target-pod" in namespace "ephemeral-containers-test-797" to be "running and ready" +Feb 25 11:12:22.367: INFO: Pod "ephemeral-containers-target-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 17.859938ms +Feb 25 11:12:22.367: INFO: The phase of Pod ephemeral-containers-target-pod is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:12:24.373: INFO: Pod "ephemeral-containers-target-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024215006s +Feb 25 11:12:24.373: INFO: The phase of Pod ephemeral-containers-target-pod is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:12:26.376: INFO: Pod "ephemeral-containers-target-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.02741974s +Feb 25 11:12:26.376: INFO: The phase of Pod ephemeral-containers-target-pod is Running (Ready = true) +Feb 25 11:12:26.376: INFO: Pod "ephemeral-containers-target-pod" satisfied condition "running and ready" +STEP: adding an ephemeral container 02/25/23 11:12:26.381 +Feb 25 11:12:26.420: INFO: Waiting up to 1m0s for pod "ephemeral-containers-target-pod" in namespace "ephemeral-containers-test-797" to be "container debugger running" +Feb 25 11:12:26.432: INFO: Pod "ephemeral-containers-target-pod": Phase="Running", Reason="", readiness=true. Elapsed: 11.474637ms +Feb 25 11:12:28.440: INFO: Pod "ephemeral-containers-target-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.019163893s +Feb 25 11:12:28.440: INFO: Pod "ephemeral-containers-target-pod" satisfied condition "container debugger running" +STEP: checking pod container endpoints 02/25/23 11:12:28.44 +Feb 25 11:12:28.440: INFO: ExecWithOptions {Command:[/bin/echo marco] Namespace:ephemeral-containers-test-797 PodName:ephemeral-containers-target-pod ContainerName:debugger Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 11:12:28.440: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 11:12:28.442: INFO: ExecWithOptions: Clientset creation +Feb 25 11:12:28.442: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/ephemeral-containers-test-797/pods/ephemeral-containers-target-pod/exec?command=%2Fbin%2Fecho&command=marco&container=debugger&container=debugger&stderr=true&stdout=true) +Feb 25 11:12:28.573: INFO: Exec stderr: "" +[AfterEach] [sig-node] Ephemeral Containers [NodeConformance] test/e2e/framework/framework.go:187 -Dec 14 16:45:14.335: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "namespaces-7563" for this suite. 12/14/22 16:45:14.343 -STEP: Destroying namespace "nsdeletetest-3234" for this suite. 12/14/22 16:45:14.362 -Dec 14 16:45:14.367: INFO: Namespace nsdeletetest-3234 was already deleted -STEP: Destroying namespace "nsdeletetest-5205" for this suite. 12/14/22 16:45:14.367 -{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all pods are removed when a namespace is deleted [Conformance]","completed":241,"skipped":4465,"failed":0} +Feb 25 11:12:28.587: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "ephemeral-containers-test-797" for this suite. 02/25/23 11:12:28.595 +{"msg":"PASSED [sig-node] Ephemeral Containers [NodeConformance] will start an ephemeral container in an existing pod [Conformance]","completed":237,"skipped":4382,"failed":0} ------------------------------ -• [SLOW TEST] [13.187 seconds] -[sig-api-machinery] Namespaces [Serial] -test/e2e/apimachinery/framework.go:23 - should ensure that all pods are removed when a namespace is deleted [Conformance] - test/e2e/apimachinery/namespace.go:242 +• [SLOW TEST] [6.315 seconds] +[sig-node] Ephemeral Containers [NodeConformance] +test/e2e/common/node/framework.go:23 + will start an ephemeral container in an existing pod [Conformance] + test/e2e/common/node/ephemeral_containers.go:45 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Namespaces [Serial] + [BeforeEach] [sig-node] Ephemeral Containers [NodeConformance] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:45:01.188 - Dec 14 16:45:01.188: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename namespaces 12/14/22 16:45:01.192 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:45:01.22 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:45:01.224 - [It] should ensure that all pods are removed when a namespace is deleted [Conformance] - test/e2e/apimachinery/namespace.go:242 - STEP: Creating a test namespace 12/14/22 16:45:01.228 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:45:01.255 - STEP: Creating a pod in the namespace 12/14/22 16:45:01.259 - STEP: Waiting for the pod to have running status 12/14/22 16:45:01.271 - Dec 14 16:45:01.271: INFO: Waiting up to 5m0s for pod "test-pod" in namespace "nsdeletetest-3234" to be "running" - Dec 14 16:45:01.275: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 3.854922ms - Dec 14 16:45:03.283: INFO: Pod "test-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.011571922s - Dec 14 16:45:03.283: INFO: Pod "test-pod" satisfied condition "running" - STEP: Deleting the namespace 12/14/22 16:45:03.283 - STEP: Waiting for the namespace to be removed. 12/14/22 16:45:03.295 - STEP: Recreating the namespace 12/14/22 16:45:14.302 - STEP: Verifying there are no pods in the namespace 12/14/22 16:45:14.329 - [AfterEach] [sig-api-machinery] Namespaces [Serial] + STEP: Creating a kubernetes client 02/25/23 11:12:22.297 + Feb 25 11:12:22.297: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename ephemeral-containers-test 02/25/23 11:12:22.301 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:12:22.322 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:12:22.328 + [BeforeEach] [sig-node] Ephemeral Containers [NodeConformance] + test/e2e/common/node/ephemeral_containers.go:38 + [It] will start an ephemeral container in an existing pod [Conformance] + test/e2e/common/node/ephemeral_containers.go:45 + STEP: creating a target pod 02/25/23 11:12:22.332 + Feb 25 11:12:22.349: INFO: Waiting up to 5m0s for pod "ephemeral-containers-target-pod" in namespace "ephemeral-containers-test-797" to be "running and ready" + Feb 25 11:12:22.367: INFO: Pod "ephemeral-containers-target-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 17.859938ms + Feb 25 11:12:22.367: INFO: The phase of Pod ephemeral-containers-target-pod is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:12:24.373: INFO: Pod "ephemeral-containers-target-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024215006s + Feb 25 11:12:24.373: INFO: The phase of Pod ephemeral-containers-target-pod is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:12:26.376: INFO: Pod "ephemeral-containers-target-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.02741974s + Feb 25 11:12:26.376: INFO: The phase of Pod ephemeral-containers-target-pod is Running (Ready = true) + Feb 25 11:12:26.376: INFO: Pod "ephemeral-containers-target-pod" satisfied condition "running and ready" + STEP: adding an ephemeral container 02/25/23 11:12:26.381 + Feb 25 11:12:26.420: INFO: Waiting up to 1m0s for pod "ephemeral-containers-target-pod" in namespace "ephemeral-containers-test-797" to be "container debugger running" + Feb 25 11:12:26.432: INFO: Pod "ephemeral-containers-target-pod": Phase="Running", Reason="", readiness=true. Elapsed: 11.474637ms + Feb 25 11:12:28.440: INFO: Pod "ephemeral-containers-target-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.019163893s + Feb 25 11:12:28.440: INFO: Pod "ephemeral-containers-target-pod" satisfied condition "container debugger running" + STEP: checking pod container endpoints 02/25/23 11:12:28.44 + Feb 25 11:12:28.440: INFO: ExecWithOptions {Command:[/bin/echo marco] Namespace:ephemeral-containers-test-797 PodName:ephemeral-containers-target-pod ContainerName:debugger Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 11:12:28.440: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 11:12:28.442: INFO: ExecWithOptions: Clientset creation + Feb 25 11:12:28.442: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/ephemeral-containers-test-797/pods/ephemeral-containers-target-pod/exec?command=%2Fbin%2Fecho&command=marco&container=debugger&container=debugger&stderr=true&stdout=true) + Feb 25 11:12:28.573: INFO: Exec stderr: "" + [AfterEach] [sig-node] Ephemeral Containers [NodeConformance] test/e2e/framework/framework.go:187 - Dec 14 16:45:14.335: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "namespaces-7563" for this suite. 12/14/22 16:45:14.343 - STEP: Destroying namespace "nsdeletetest-3234" for this suite. 12/14/22 16:45:14.362 - Dec 14 16:45:14.367: INFO: Namespace nsdeletetest-3234 was already deleted - STEP: Destroying namespace "nsdeletetest-5205" for this suite. 12/14/22 16:45:14.367 + Feb 25 11:12:28.587: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "ephemeral-containers-test-797" for this suite. 02/25/23 11:12:28.595 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SS ------------------------------ -[sig-storage] ConfigMap - should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:108 -[BeforeEach] [sig-storage] ConfigMap +[sig-node] Container Runtime blackbox test on terminated container + should report termination message from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:247 +[BeforeEach] [sig-node] Container Runtime test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:45:14.383 -Dec 14 16:45:14.383: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename configmap 12/14/22 16:45:14.391 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:45:14.424 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:45:14.427 -[It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:108 -STEP: Creating configMap with name configmap-test-volume-map-63b053bf-2889-4514-a6ae-0a501e354db8 12/14/22 16:45:14.43 -STEP: Creating a pod to test consume configMaps 12/14/22 16:45:14.437 -Dec 14 16:45:14.447: INFO: Waiting up to 5m0s for pod "pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39" in namespace "configmap-449" to be "Succeeded or Failed" -Dec 14 16:45:14.454: INFO: Pod "pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39": Phase="Pending", Reason="", readiness=false. Elapsed: 6.471082ms -Dec 14 16:45:16.460: INFO: Pod "pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013111524s -Dec 14 16:45:18.466: INFO: Pod "pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018856613s -STEP: Saw pod success 12/14/22 16:45:18.466 -Dec 14 16:45:18.466: INFO: Pod "pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39" satisfied condition "Succeeded or Failed" -Dec 14 16:45:18.472: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39 container agnhost-container: -STEP: delete the pod 12/14/22 16:45:18.481 -Dec 14 16:45:18.500: INFO: Waiting for pod pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39 to disappear -Dec 14 16:45:18.504: INFO: Pod pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39 no longer exists -[AfterEach] [sig-storage] ConfigMap +STEP: Creating a kubernetes client 02/25/23 11:12:28.614 +Feb 25 11:12:28.615: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-runtime 02/25/23 11:12:28.617 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:12:28.644 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:12:28.65 +[It] should report termination message from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:247 +STEP: create the container 02/25/23 11:12:28.655 +STEP: wait for the container to reach Succeeded 02/25/23 11:12:28.668 +STEP: get the container status 02/25/23 11:12:33.727 +STEP: the container should be terminated 02/25/23 11:12:33.734 +STEP: the termination message should be set 02/25/23 11:12:33.734 +Feb 25 11:12:33.734: INFO: Expected: &{OK} to match Container's Termination Message: OK -- +STEP: delete the container 02/25/23 11:12:33.734 +[AfterEach] [sig-node] Container Runtime test/e2e/framework/framework.go:187 -Dec 14 16:45:18.505: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-449" for this suite. 12/14/22 16:45:18.514 -{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance]","completed":242,"skipped":4501,"failed":0} +Feb 25 11:12:33.762: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-runtime-7689" for this suite. 02/25/23 11:12:33.774 +{"msg":"PASSED [sig-node] Container Runtime blackbox test on terminated container should report termination message from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","completed":238,"skipped":4384,"failed":0} ------------------------------ -• [4.141 seconds] -[sig-storage] ConfigMap -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:108 +• [SLOW TEST] [5.178 seconds] +[sig-node] Container Runtime +test/e2e/common/node/framework.go:23 + blackbox test + test/e2e/common/node/runtime.go:43 + on terminated container + test/e2e/common/node/runtime.go:136 + should report termination message from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:247 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] ConfigMap + [BeforeEach] [sig-node] Container Runtime test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:45:14.383 - Dec 14 16:45:14.383: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename configmap 12/14/22 16:45:14.391 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:45:14.424 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:45:14.427 - [It] should be consumable from pods in volume with mappings as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:108 - STEP: Creating configMap with name configmap-test-volume-map-63b053bf-2889-4514-a6ae-0a501e354db8 12/14/22 16:45:14.43 - STEP: Creating a pod to test consume configMaps 12/14/22 16:45:14.437 - Dec 14 16:45:14.447: INFO: Waiting up to 5m0s for pod "pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39" in namespace "configmap-449" to be "Succeeded or Failed" - Dec 14 16:45:14.454: INFO: Pod "pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39": Phase="Pending", Reason="", readiness=false. Elapsed: 6.471082ms - Dec 14 16:45:16.460: INFO: Pod "pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013111524s - Dec 14 16:45:18.466: INFO: Pod "pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018856613s - STEP: Saw pod success 12/14/22 16:45:18.466 - Dec 14 16:45:18.466: INFO: Pod "pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39" satisfied condition "Succeeded or Failed" - Dec 14 16:45:18.472: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39 container agnhost-container: - STEP: delete the pod 12/14/22 16:45:18.481 - Dec 14 16:45:18.500: INFO: Waiting for pod pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39 to disappear - Dec 14 16:45:18.504: INFO: Pod pod-configmaps-02d8f195-ba12-41a7-8da7-768cd914fa39 no longer exists - [AfterEach] [sig-storage] ConfigMap + STEP: Creating a kubernetes client 02/25/23 11:12:28.614 + Feb 25 11:12:28.615: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-runtime 02/25/23 11:12:28.617 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:12:28.644 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:12:28.65 + [It] should report termination message from file when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:247 + STEP: create the container 02/25/23 11:12:28.655 + STEP: wait for the container to reach Succeeded 02/25/23 11:12:28.668 + STEP: get the container status 02/25/23 11:12:33.727 + STEP: the container should be terminated 02/25/23 11:12:33.734 + STEP: the termination message should be set 02/25/23 11:12:33.734 + Feb 25 11:12:33.734: INFO: Expected: &{OK} to match Container's Termination Message: OK -- + STEP: delete the container 02/25/23 11:12:33.734 + [AfterEach] [sig-node] Container Runtime test/e2e/framework/framework.go:187 - Dec 14 16:45:18.505: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "configmap-449" for this suite. 12/14/22 16:45:18.514 + Feb 25 11:12:33.762: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-runtime-7689" for this suite. 02/25/23 11:12:33.774 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] Garbage collector - should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] - test/e2e/apimachinery/garbage_collector.go:650 -[BeforeEach] [sig-api-machinery] Garbage collector +[sig-api-machinery] ResourceQuota + should create a ResourceQuota and capture the life of a replication controller. [Conformance] + test/e2e/apimachinery/resource_quota.go:382 +[BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:45:18.534 -Dec 14 16:45:18.534: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename gc 12/14/22 16:45:18.536 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:45:18.579 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:45:18.584 -[It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] - test/e2e/apimachinery/garbage_collector.go:650 -STEP: create the rc 12/14/22 16:45:18.599 -STEP: delete the rc 12/14/22 16:45:23.646 -STEP: wait for the rc to be deleted 12/14/22 16:45:23.692 -Dec 14 16:45:24.806: INFO: 82 pods remaining -Dec 14 16:45:24.808: INFO: 80 pods has nil DeletionTimestamp -Dec 14 16:45:24.809: INFO: -Dec 14 16:45:25.762: INFO: 73 pods remaining -Dec 14 16:45:25.763: INFO: 72 pods has nil DeletionTimestamp -Dec 14 16:45:25.763: INFO: -Dec 14 16:45:26.822: INFO: 58 pods remaining -Dec 14 16:45:26.823: INFO: 58 pods has nil DeletionTimestamp -Dec 14 16:45:26.823: INFO: -Dec 14 16:45:27.716: INFO: 48 pods remaining -Dec 14 16:45:27.716: INFO: 42 pods has nil DeletionTimestamp -Dec 14 16:45:27.716: INFO: -Dec 14 16:45:28.720: INFO: 30 pods remaining -Dec 14 16:45:28.720: INFO: 30 pods has nil DeletionTimestamp -Dec 14 16:45:28.720: INFO: -Dec 14 16:45:29.767: INFO: 18 pods remaining -Dec 14 16:45:29.767: INFO: 17 pods has nil DeletionTimestamp -Dec 14 16:45:29.767: INFO: -Dec 14 16:45:30.755: INFO: 0 pods remaining -Dec 14 16:45:30.755: INFO: 0 pods has nil DeletionTimestamp -Dec 14 16:45:30.755: INFO: -STEP: Gathering metrics 12/14/22 16:45:31.705 -Dec 14 16:45:32.271: INFO: Waiting up to 5m0s for pod "kube-controller-manager-kiedai4iakae-2" in namespace "kube-system" to be "running and ready" -Dec 14 16:45:32.291: INFO: Pod "kube-controller-manager-kiedai4iakae-2": Phase="Running", Reason="", readiness=true. Elapsed: 19.670494ms -Dec 14 16:45:32.291: INFO: The phase of Pod kube-controller-manager-kiedai4iakae-2 is Running (Ready = true) -Dec 14 16:45:32.291: INFO: Pod "kube-controller-manager-kiedai4iakae-2" satisfied condition "running and ready" -Dec 14 16:45:32.828: INFO: For apiserver_request_total: -For apiserver_request_latency_seconds: -For apiserver_init_events_total: -For garbage_collector_attempt_to_delete_queue_latency: -For garbage_collector_attempt_to_delete_work_duration: -For garbage_collector_attempt_to_orphan_queue_latency: -For garbage_collector_attempt_to_orphan_work_duration: -For garbage_collector_dirty_processing_latency_microseconds: -For garbage_collector_event_processing_latency_microseconds: -For garbage_collector_graph_changes_queue_latency: -For garbage_collector_graph_changes_work_duration: -For garbage_collector_orphan_processing_latency_microseconds: -For namespace_queue_latency: -For namespace_queue_latency_sum: -For namespace_queue_latency_count: -For namespace_retries: -For namespace_work_duration: -For namespace_work_duration_sum: -For namespace_work_duration_count: -For function_duration_seconds: -For errors_total: -For evicted_pods_total: - -[AfterEach] [sig-api-machinery] Garbage collector +STEP: Creating a kubernetes client 02/25/23 11:12:33.799 +Feb 25 11:12:33.799: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename resourcequota 02/25/23 11:12:33.805 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:12:33.846 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:12:33.859 +[It] should create a ResourceQuota and capture the life of a replication controller. [Conformance] + test/e2e/apimachinery/resource_quota.go:382 +STEP: Counting existing ResourceQuota 02/25/23 11:12:33.866 +STEP: Creating a ResourceQuota 02/25/23 11:12:38.886 +STEP: Ensuring resource quota status is calculated 02/25/23 11:12:38.906 +STEP: Creating a ReplicationController 02/25/23 11:12:40.915 +STEP: Ensuring resource quota status captures replication controller creation 02/25/23 11:12:40.957 +STEP: Deleting a ReplicationController 02/25/23 11:12:42.965 +STEP: Ensuring resource quota status released usage 02/25/23 11:12:42.976 +[AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 -Dec 14 16:45:32.828: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "gc-4821" for this suite. 12/14/22 16:45:32.835 -{"msg":"PASSED [sig-api-machinery] Garbage collector should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance]","completed":243,"skipped":4540,"failed":0} +Feb 25 11:12:44.986: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-7543" for this suite. 02/25/23 11:12:45.001 +{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replication controller. [Conformance]","completed":239,"skipped":4442,"failed":0} ------------------------------ -• [SLOW TEST] [14.317 seconds] -[sig-api-machinery] Garbage collector +• [SLOW TEST] [11.231 seconds] +[sig-api-machinery] ResourceQuota test/e2e/apimachinery/framework.go:23 - should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] - test/e2e/apimachinery/garbage_collector.go:650 + should create a ResourceQuota and capture the life of a replication controller. [Conformance] + test/e2e/apimachinery/resource_quota.go:382 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Garbage collector + [BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:45:18.534 - Dec 14 16:45:18.534: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename gc 12/14/22 16:45:18.536 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:45:18.579 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:45:18.584 - [It] should keep the rc around until all its pods are deleted if the deleteOptions says so [Conformance] - test/e2e/apimachinery/garbage_collector.go:650 - STEP: create the rc 12/14/22 16:45:18.599 - STEP: delete the rc 12/14/22 16:45:23.646 - STEP: wait for the rc to be deleted 12/14/22 16:45:23.692 - Dec 14 16:45:24.806: INFO: 82 pods remaining - Dec 14 16:45:24.808: INFO: 80 pods has nil DeletionTimestamp - Dec 14 16:45:24.809: INFO: - Dec 14 16:45:25.762: INFO: 73 pods remaining - Dec 14 16:45:25.763: INFO: 72 pods has nil DeletionTimestamp - Dec 14 16:45:25.763: INFO: - Dec 14 16:45:26.822: INFO: 58 pods remaining - Dec 14 16:45:26.823: INFO: 58 pods has nil DeletionTimestamp - Dec 14 16:45:26.823: INFO: - Dec 14 16:45:27.716: INFO: 48 pods remaining - Dec 14 16:45:27.716: INFO: 42 pods has nil DeletionTimestamp - Dec 14 16:45:27.716: INFO: - Dec 14 16:45:28.720: INFO: 30 pods remaining - Dec 14 16:45:28.720: INFO: 30 pods has nil DeletionTimestamp - Dec 14 16:45:28.720: INFO: - Dec 14 16:45:29.767: INFO: 18 pods remaining - Dec 14 16:45:29.767: INFO: 17 pods has nil DeletionTimestamp - Dec 14 16:45:29.767: INFO: - Dec 14 16:45:30.755: INFO: 0 pods remaining - Dec 14 16:45:30.755: INFO: 0 pods has nil DeletionTimestamp - Dec 14 16:45:30.755: INFO: - STEP: Gathering metrics 12/14/22 16:45:31.705 - Dec 14 16:45:32.271: INFO: Waiting up to 5m0s for pod "kube-controller-manager-kiedai4iakae-2" in namespace "kube-system" to be "running and ready" - Dec 14 16:45:32.291: INFO: Pod "kube-controller-manager-kiedai4iakae-2": Phase="Running", Reason="", readiness=true. Elapsed: 19.670494ms - Dec 14 16:45:32.291: INFO: The phase of Pod kube-controller-manager-kiedai4iakae-2 is Running (Ready = true) - Dec 14 16:45:32.291: INFO: Pod "kube-controller-manager-kiedai4iakae-2" satisfied condition "running and ready" - Dec 14 16:45:32.828: INFO: For apiserver_request_total: - For apiserver_request_latency_seconds: - For apiserver_init_events_total: - For garbage_collector_attempt_to_delete_queue_latency: - For garbage_collector_attempt_to_delete_work_duration: - For garbage_collector_attempt_to_orphan_queue_latency: - For garbage_collector_attempt_to_orphan_work_duration: - For garbage_collector_dirty_processing_latency_microseconds: - For garbage_collector_event_processing_latency_microseconds: - For garbage_collector_graph_changes_queue_latency: - For garbage_collector_graph_changes_work_duration: - For garbage_collector_orphan_processing_latency_microseconds: - For namespace_queue_latency: - For namespace_queue_latency_sum: - For namespace_queue_latency_count: - For namespace_retries: - For namespace_work_duration: - For namespace_work_duration_sum: - For namespace_work_duration_count: - For function_duration_seconds: - For errors_total: - For evicted_pods_total: - - [AfterEach] [sig-api-machinery] Garbage collector + STEP: Creating a kubernetes client 02/25/23 11:12:33.799 + Feb 25 11:12:33.799: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename resourcequota 02/25/23 11:12:33.805 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:12:33.846 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:12:33.859 + [It] should create a ResourceQuota and capture the life of a replication controller. [Conformance] + test/e2e/apimachinery/resource_quota.go:382 + STEP: Counting existing ResourceQuota 02/25/23 11:12:33.866 + STEP: Creating a ResourceQuota 02/25/23 11:12:38.886 + STEP: Ensuring resource quota status is calculated 02/25/23 11:12:38.906 + STEP: Creating a ReplicationController 02/25/23 11:12:40.915 + STEP: Ensuring resource quota status captures replication controller creation 02/25/23 11:12:40.957 + STEP: Deleting a ReplicationController 02/25/23 11:12:42.965 + STEP: Ensuring resource quota status released usage 02/25/23 11:12:42.976 + [AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 - Dec 14 16:45:32.828: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "gc-4821" for this suite. 12/14/22 16:45:32.835 + Feb 25 11:12:44.986: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "resourcequota-7543" for this suite. 02/25/23 11:12:45.001 << End Captured GinkgoWriter Output ------------------------------ -S ------------------------------- -[sig-cli] Kubectl client Kubectl server-side dry-run - should check if kubectl can dry-run update Pods [Conformance] - test/e2e/kubectl/kubectl.go:960 -[BeforeEach] [sig-cli] Kubectl client +[sig-network] Services + should have session affinity work for NodePort service [LinuxOnly] [Conformance] + test/e2e/network/service.go:2204 +[BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:45:32.852 -Dec 14 16:45:32.853: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 16:45:32.86 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:45:32.911 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:45:32.915 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 -[It] should check if kubectl can dry-run update Pods [Conformance] - test/e2e/kubectl/kubectl.go:960 -STEP: running the image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 12/14/22 16:45:32.918 -Dec 14 16:45:32.921: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-5410 run e2e-test-httpd-pod --image=registry.k8s.io/e2e-test-images/httpd:2.4.38-2 --pod-running-timeout=2m0s --labels=run=e2e-test-httpd-pod' -Dec 14 16:45:33.165: INFO: stderr: "" -Dec 14 16:45:33.165: INFO: stdout: "pod/e2e-test-httpd-pod created\n" -STEP: replace the image in the pod with server-side dry-run 12/14/22 16:45:33.165 -Dec 14 16:45:33.166: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-5410 patch pod e2e-test-httpd-pod -p {"spec":{"containers":[{"name": "e2e-test-httpd-pod","image": "registry.k8s.io/e2e-test-images/busybox:1.29-2"}]}} --dry-run=server' -Dec 14 16:45:34.667: INFO: stderr: "" -Dec 14 16:45:34.667: INFO: stdout: "pod/e2e-test-httpd-pod patched\n" -STEP: verifying the pod e2e-test-httpd-pod has the right image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 12/14/22 16:45:34.667 -Dec 14 16:45:34.676: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-5410 delete pods e2e-test-httpd-pod' -Dec 14 16:46:11.621: INFO: stderr: "" -Dec 14 16:46:11.621: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" -[AfterEach] [sig-cli] Kubectl client +STEP: Creating a kubernetes client 02/25/23 11:12:45.036 +Feb 25 11:12:45.037: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 11:12:45.039 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:12:45.077 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:12:45.089 +[BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 +[It] should have session affinity work for NodePort service [LinuxOnly] [Conformance] + test/e2e/network/service.go:2204 +STEP: creating service in namespace services-6864 02/25/23 11:12:45.095 +STEP: creating service affinity-nodeport in namespace services-6864 02/25/23 11:12:45.098 +STEP: creating replication controller affinity-nodeport in namespace services-6864 02/25/23 11:12:45.265 +I0225 11:12:45.284999 13 runners.go:193] Created replication controller with name: affinity-nodeport, namespace: services-6864, replica count: 3 +I0225 11:12:48.336586 13 runners.go:193] affinity-nodeport Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Feb 25 11:12:48.359: INFO: Creating new exec pod +Feb 25 11:12:48.370: INFO: Waiting up to 5m0s for pod "execpod-affinityjhqlc" in namespace "services-6864" to be "running" +Feb 25 11:12:48.377: INFO: Pod "execpod-affinityjhqlc": Phase="Pending", Reason="", readiness=false. Elapsed: 6.835117ms +Feb 25 11:12:50.382: INFO: Pod "execpod-affinityjhqlc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012539118s +Feb 25 11:12:52.386: INFO: Pod "execpod-affinityjhqlc": Phase="Running", Reason="", readiness=true. Elapsed: 4.015829906s +Feb 25 11:12:52.386: INFO: Pod "execpod-affinityjhqlc" satisfied condition "running" +Feb 25 11:12:53.404: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6864 exec execpod-affinityjhqlc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-nodeport 80' +Feb 25 11:12:53.828: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-nodeport 80\nConnection to affinity-nodeport 80 port [tcp/http] succeeded!\n" +Feb 25 11:12:53.828: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:12:53.832: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6864 exec execpod-affinityjhqlc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.19.111 80' +Feb 25 11:12:54.381: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.19.111 80\nConnection to 10.233.19.111 80 port [tcp/http] succeeded!\n" +Feb 25 11:12:54.381: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:12:54.381: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6864 exec execpod-affinityjhqlc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.117 32000' +Feb 25 11:12:54.685: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.117 32000\nConnection to 192.168.121.117 32000 port [tcp/*] succeeded!\n" +Feb 25 11:12:54.685: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:12:54.686: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6864 exec execpod-affinityjhqlc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.250 32000' +Feb 25 11:12:55.027: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.250 32000\nConnection to 192.168.121.250 32000 port [tcp/*] succeeded!\n" +Feb 25 11:12:55.027: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:12:55.028: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6864 exec execpod-affinityjhqlc -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.250:32000/ ; done' +Feb 25 11:12:55.598: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n" +Feb 25 11:12:55.598: INFO: stdout: "\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4" +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 +Feb 25 11:12:55.599: INFO: Cleaning up the exec pod +STEP: deleting ReplicationController affinity-nodeport in namespace services-6864, will wait for the garbage collector to delete the pods 02/25/23 11:12:55.639 +Feb 25 11:12:55.731: INFO: Deleting ReplicationController affinity-nodeport took: 32.112049ms +Feb 25 11:12:55.832: INFO: Terminating ReplicationController affinity-nodeport pods took: 100.918162ms +[AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 16:46:11.622: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-5410" for this suite. 12/14/22 16:46:11.63 -{"msg":"PASSED [sig-cli] Kubectl client Kubectl server-side dry-run should check if kubectl can dry-run update Pods [Conformance]","completed":244,"skipped":4541,"failed":0} +Feb 25 11:12:59.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-6864" for this suite. 02/25/23 11:12:59.243 +[AfterEach] [sig-network] Services + test/e2e/network/service.go:762 +{"msg":"PASSED [sig-network] Services should have session affinity work for NodePort service [LinuxOnly] [Conformance]","completed":240,"skipped":4442,"failed":0} ------------------------------ -• [SLOW TEST] [38.789 seconds] -[sig-cli] Kubectl client -test/e2e/kubectl/framework.go:23 - Kubectl server-side dry-run - test/e2e/kubectl/kubectl.go:954 - should check if kubectl can dry-run update Pods [Conformance] - test/e2e/kubectl/kubectl.go:960 +• [SLOW TEST] [14.219 seconds] +[sig-network] Services +test/e2e/network/common/framework.go:23 + should have session affinity work for NodePort service [LinuxOnly] [Conformance] + test/e2e/network/service.go:2204 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-cli] Kubectl client + [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:45:32.852 - Dec 14 16:45:32.853: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 16:45:32.86 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:45:32.911 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:45:32.915 - [BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 - [It] should check if kubectl can dry-run update Pods [Conformance] - test/e2e/kubectl/kubectl.go:960 - STEP: running the image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 12/14/22 16:45:32.918 - Dec 14 16:45:32.921: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-5410 run e2e-test-httpd-pod --image=registry.k8s.io/e2e-test-images/httpd:2.4.38-2 --pod-running-timeout=2m0s --labels=run=e2e-test-httpd-pod' - Dec 14 16:45:33.165: INFO: stderr: "" - Dec 14 16:45:33.165: INFO: stdout: "pod/e2e-test-httpd-pod created\n" - STEP: replace the image in the pod with server-side dry-run 12/14/22 16:45:33.165 - Dec 14 16:45:33.166: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-5410 patch pod e2e-test-httpd-pod -p {"spec":{"containers":[{"name": "e2e-test-httpd-pod","image": "registry.k8s.io/e2e-test-images/busybox:1.29-2"}]}} --dry-run=server' - Dec 14 16:45:34.667: INFO: stderr: "" - Dec 14 16:45:34.667: INFO: stdout: "pod/e2e-test-httpd-pod patched\n" - STEP: verifying the pod e2e-test-httpd-pod has the right image registry.k8s.io/e2e-test-images/httpd:2.4.38-2 12/14/22 16:45:34.667 - Dec 14 16:45:34.676: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-5410 delete pods e2e-test-httpd-pod' - Dec 14 16:46:11.621: INFO: stderr: "" - Dec 14 16:46:11.621: INFO: stdout: "pod \"e2e-test-httpd-pod\" deleted\n" - [AfterEach] [sig-cli] Kubectl client + STEP: Creating a kubernetes client 02/25/23 11:12:45.036 + Feb 25 11:12:45.037: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 11:12:45.039 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:12:45.077 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:12:45.089 + [BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 + [It] should have session affinity work for NodePort service [LinuxOnly] [Conformance] + test/e2e/network/service.go:2204 + STEP: creating service in namespace services-6864 02/25/23 11:12:45.095 + STEP: creating service affinity-nodeport in namespace services-6864 02/25/23 11:12:45.098 + STEP: creating replication controller affinity-nodeport in namespace services-6864 02/25/23 11:12:45.265 + I0225 11:12:45.284999 13 runners.go:193] Created replication controller with name: affinity-nodeport, namespace: services-6864, replica count: 3 + I0225 11:12:48.336586 13 runners.go:193] affinity-nodeport Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + Feb 25 11:12:48.359: INFO: Creating new exec pod + Feb 25 11:12:48.370: INFO: Waiting up to 5m0s for pod "execpod-affinityjhqlc" in namespace "services-6864" to be "running" + Feb 25 11:12:48.377: INFO: Pod "execpod-affinityjhqlc": Phase="Pending", Reason="", readiness=false. Elapsed: 6.835117ms + Feb 25 11:12:50.382: INFO: Pod "execpod-affinityjhqlc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012539118s + Feb 25 11:12:52.386: INFO: Pod "execpod-affinityjhqlc": Phase="Running", Reason="", readiness=true. Elapsed: 4.015829906s + Feb 25 11:12:52.386: INFO: Pod "execpod-affinityjhqlc" satisfied condition "running" + Feb 25 11:12:53.404: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6864 exec execpod-affinityjhqlc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-nodeport 80' + Feb 25 11:12:53.828: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-nodeport 80\nConnection to affinity-nodeport 80 port [tcp/http] succeeded!\n" + Feb 25 11:12:53.828: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:12:53.832: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6864 exec execpod-affinityjhqlc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.19.111 80' + Feb 25 11:12:54.381: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.19.111 80\nConnection to 10.233.19.111 80 port [tcp/http] succeeded!\n" + Feb 25 11:12:54.381: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:12:54.381: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6864 exec execpod-affinityjhqlc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.117 32000' + Feb 25 11:12:54.685: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.117 32000\nConnection to 192.168.121.117 32000 port [tcp/*] succeeded!\n" + Feb 25 11:12:54.685: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:12:54.686: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6864 exec execpod-affinityjhqlc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.250 32000' + Feb 25 11:12:55.027: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.250 32000\nConnection to 192.168.121.250 32000 port [tcp/*] succeeded!\n" + Feb 25 11:12:55.027: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:12:55.028: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6864 exec execpod-affinityjhqlc -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.250:32000/ ; done' + Feb 25 11:12:55.598: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:32000/\n" + Feb 25 11:12:55.598: INFO: stdout: "\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4\naffinity-nodeport-dx8h4" + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.598: INFO: Received response from host: affinity-nodeport-dx8h4 + Feb 25 11:12:55.599: INFO: Cleaning up the exec pod + STEP: deleting ReplicationController affinity-nodeport in namespace services-6864, will wait for the garbage collector to delete the pods 02/25/23 11:12:55.639 + Feb 25 11:12:55.731: INFO: Deleting ReplicationController affinity-nodeport took: 32.112049ms + Feb 25 11:12:55.832: INFO: Terminating ReplicationController affinity-nodeport pods took: 100.918162ms + [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 16:46:11.622: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-5410" for this suite. 12/14/22 16:46:11.63 + Feb 25 11:12:59.230: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-6864" for this suite. 02/25/23 11:12:59.243 + [AfterEach] [sig-network] Services + test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -SSSSS +SSSSSSSSSSSSS ------------------------------ -[sig-storage] Projected secret - optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:214 -[BeforeEach] [sig-storage] Projected secret +[sig-network] Services + should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance] + test/e2e/network/service.go:2237 +[BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:46:11.646 -Dec 14 16:46:11.647: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 16:46:11.649 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:46:11.678 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:46:11.681 -[It] optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:214 -STEP: Creating secret with name s-test-opt-del-b8f1994e-de66-4aeb-b3c7-81d60f8ac5e4 12/14/22 16:46:11.692 -STEP: Creating secret with name s-test-opt-upd-11818807-70ab-464e-ae42-f6e355f7a99b 12/14/22 16:46:11.699 -STEP: Creating the pod 12/14/22 16:46:11.708 -Dec 14 16:46:11.722: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-65681b35-6188-475e-98f0-b63fff189585" in namespace "projected-7807" to be "running and ready" -Dec 14 16:46:11.725: INFO: Pod "pod-projected-secrets-65681b35-6188-475e-98f0-b63fff189585": Phase="Pending", Reason="", readiness=false. Elapsed: 3.357707ms -Dec 14 16:46:11.725: INFO: The phase of Pod pod-projected-secrets-65681b35-6188-475e-98f0-b63fff189585 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:46:13.734: INFO: Pod "pod-projected-secrets-65681b35-6188-475e-98f0-b63fff189585": Phase="Running", Reason="", readiness=true. Elapsed: 2.012707775s -Dec 14 16:46:13.735: INFO: The phase of Pod pod-projected-secrets-65681b35-6188-475e-98f0-b63fff189585 is Running (Ready = true) -Dec 14 16:46:13.735: INFO: Pod "pod-projected-secrets-65681b35-6188-475e-98f0-b63fff189585" satisfied condition "running and ready" -STEP: Deleting secret s-test-opt-del-b8f1994e-de66-4aeb-b3c7-81d60f8ac5e4 12/14/22 16:46:14.331 -STEP: Updating secret s-test-opt-upd-11818807-70ab-464e-ae42-f6e355f7a99b 12/14/22 16:46:14.34 -STEP: Creating secret with name s-test-opt-create-7e9de2e5-4dc7-4c7c-aeb4-638e0f37b3fe 12/14/22 16:46:14.349 -STEP: waiting to observe update in volume 12/14/22 16:46:14.355 -[AfterEach] [sig-storage] Projected secret +STEP: Creating a kubernetes client 02/25/23 11:12:59.261 +Feb 25 11:12:59.261: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 11:12:59.266 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:12:59.326 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:12:59.332 +[BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 +[It] should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance] + test/e2e/network/service.go:2237 +STEP: creating service in namespace services-9693 02/25/23 11:12:59.336 +STEP: creating service affinity-nodeport-transition in namespace services-9693 02/25/23 11:12:59.336 +STEP: creating replication controller affinity-nodeport-transition in namespace services-9693 02/25/23 11:12:59.372 +I0225 11:12:59.405153 13 runners.go:193] Created replication controller with name: affinity-nodeport-transition, namespace: services-9693, replica count: 3 +I0225 11:13:02.457238 13 runners.go:193] affinity-nodeport-transition Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Feb 25 11:13:02.485: INFO: Creating new exec pod +Feb 25 11:13:02.509: INFO: Waiting up to 5m0s for pod "execpod-affinityr8q9l" in namespace "services-9693" to be "running" +Feb 25 11:13:02.531: INFO: Pod "execpod-affinityr8q9l": Phase="Pending", Reason="", readiness=false. Elapsed: 21.641051ms +Feb 25 11:13:04.542: INFO: Pod "execpod-affinityr8q9l": Phase="Pending", Reason="", readiness=false. Elapsed: 2.032982746s +Feb 25 11:13:06.540: INFO: Pod "execpod-affinityr8q9l": Phase="Running", Reason="", readiness=true. Elapsed: 4.031516055s +Feb 25 11:13:06.541: INFO: Pod "execpod-affinityr8q9l" satisfied condition "running" +Feb 25 11:13:07.551: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9693 exec execpod-affinityr8q9l -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-nodeport-transition 80' +Feb 25 11:13:08.178: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-nodeport-transition 80\nConnection to affinity-nodeport-transition 80 port [tcp/http] succeeded!\n" +Feb 25 11:13:08.178: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:13:08.181: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9693 exec execpod-affinityr8q9l -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.37.115 80' +Feb 25 11:13:08.535: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.37.115 80\nConnection to 10.233.37.115 80 port [tcp/http] succeeded!\n" +Feb 25 11:13:08.535: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:13:08.535: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9693 exec execpod-affinityr8q9l -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.250 30798' +Feb 25 11:13:08.834: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.250 30798\nConnection to 192.168.121.250 30798 port [tcp/*] succeeded!\n" +Feb 25 11:13:08.834: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:13:08.835: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9693 exec execpod-affinityr8q9l -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.44 30798' +Feb 25 11:13:09.177: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.44 30798\nConnection to 192.168.121.44 30798 port [tcp/*] succeeded!\n" +Feb 25 11:13:09.177: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:13:09.318: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9693 exec execpod-affinityr8q9l -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.250:30798/ ; done' +Feb 25 11:13:09.853: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n" +Feb 25 11:13:09.854: INFO: stdout: "\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-pqxxm\naffinity-nodeport-transition-pqxxm\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-zk9lq\naffinity-nodeport-transition-zk9lq\naffinity-nodeport-transition-pqxxm\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-pqxxm\naffinity-nodeport-transition-pqxxm\naffinity-nodeport-transition-zk9lq\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn" +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-pqxxm +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-pqxxm +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-zk9lq +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-zk9lq +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-pqxxm +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-pqxxm +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-pqxxm +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-zk9lq +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:09.870: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9693 exec execpod-affinityr8q9l -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.250:30798/ ; done' +Feb 25 11:13:10.782: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n" +Feb 25 11:13:10.782: INFO: stdout: "\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn" +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn +Feb 25 11:13:10.783: INFO: Cleaning up the exec pod +STEP: deleting ReplicationController affinity-nodeport-transition in namespace services-9693, will wait for the garbage collector to delete the pods 02/25/23 11:13:10.814 +Feb 25 11:13:10.897: INFO: Deleting ReplicationController affinity-nodeport-transition took: 20.717446ms +Feb 25 11:13:10.998: INFO: Terminating ReplicationController affinity-nodeport-transition pods took: 100.539817ms +[AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 16:46:16.398: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-7807" for this suite. 12/14/22 16:46:16.405 -{"msg":"PASSED [sig-storage] Projected secret optional updates should be reflected in volume [NodeConformance] [Conformance]","completed":245,"skipped":4546,"failed":0} +Feb 25 11:13:14.237: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-9693" for this suite. 02/25/23 11:13:14.254 +[AfterEach] [sig-network] Services + test/e2e/network/service.go:762 +{"msg":"PASSED [sig-network] Services should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance]","completed":241,"skipped":4455,"failed":0} ------------------------------ -• [4.770 seconds] -[sig-storage] Projected secret -test/e2e/common/storage/framework.go:23 - optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:214 +• [SLOW TEST] [15.010 seconds] +[sig-network] Services +test/e2e/network/common/framework.go:23 + should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance] + test/e2e/network/service.go:2237 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected secret + [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:46:11.646 - Dec 14 16:46:11.647: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 16:46:11.649 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:46:11.678 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:46:11.681 - [It] optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:214 - STEP: Creating secret with name s-test-opt-del-b8f1994e-de66-4aeb-b3c7-81d60f8ac5e4 12/14/22 16:46:11.692 - STEP: Creating secret with name s-test-opt-upd-11818807-70ab-464e-ae42-f6e355f7a99b 12/14/22 16:46:11.699 - STEP: Creating the pod 12/14/22 16:46:11.708 - Dec 14 16:46:11.722: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-65681b35-6188-475e-98f0-b63fff189585" in namespace "projected-7807" to be "running and ready" - Dec 14 16:46:11.725: INFO: Pod "pod-projected-secrets-65681b35-6188-475e-98f0-b63fff189585": Phase="Pending", Reason="", readiness=false. Elapsed: 3.357707ms - Dec 14 16:46:11.725: INFO: The phase of Pod pod-projected-secrets-65681b35-6188-475e-98f0-b63fff189585 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:46:13.734: INFO: Pod "pod-projected-secrets-65681b35-6188-475e-98f0-b63fff189585": Phase="Running", Reason="", readiness=true. Elapsed: 2.012707775s - Dec 14 16:46:13.735: INFO: The phase of Pod pod-projected-secrets-65681b35-6188-475e-98f0-b63fff189585 is Running (Ready = true) - Dec 14 16:46:13.735: INFO: Pod "pod-projected-secrets-65681b35-6188-475e-98f0-b63fff189585" satisfied condition "running and ready" - STEP: Deleting secret s-test-opt-del-b8f1994e-de66-4aeb-b3c7-81d60f8ac5e4 12/14/22 16:46:14.331 - STEP: Updating secret s-test-opt-upd-11818807-70ab-464e-ae42-f6e355f7a99b 12/14/22 16:46:14.34 - STEP: Creating secret with name s-test-opt-create-7e9de2e5-4dc7-4c7c-aeb4-638e0f37b3fe 12/14/22 16:46:14.349 - STEP: waiting to observe update in volume 12/14/22 16:46:14.355 - [AfterEach] [sig-storage] Projected secret + STEP: Creating a kubernetes client 02/25/23 11:12:59.261 + Feb 25 11:12:59.261: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 11:12:59.266 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:12:59.326 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:12:59.332 + [BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 + [It] should be able to switch session affinity for NodePort service [LinuxOnly] [Conformance] + test/e2e/network/service.go:2237 + STEP: creating service in namespace services-9693 02/25/23 11:12:59.336 + STEP: creating service affinity-nodeport-transition in namespace services-9693 02/25/23 11:12:59.336 + STEP: creating replication controller affinity-nodeport-transition in namespace services-9693 02/25/23 11:12:59.372 + I0225 11:12:59.405153 13 runners.go:193] Created replication controller with name: affinity-nodeport-transition, namespace: services-9693, replica count: 3 + I0225 11:13:02.457238 13 runners.go:193] affinity-nodeport-transition Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + Feb 25 11:13:02.485: INFO: Creating new exec pod + Feb 25 11:13:02.509: INFO: Waiting up to 5m0s for pod "execpod-affinityr8q9l" in namespace "services-9693" to be "running" + Feb 25 11:13:02.531: INFO: Pod "execpod-affinityr8q9l": Phase="Pending", Reason="", readiness=false. Elapsed: 21.641051ms + Feb 25 11:13:04.542: INFO: Pod "execpod-affinityr8q9l": Phase="Pending", Reason="", readiness=false. Elapsed: 2.032982746s + Feb 25 11:13:06.540: INFO: Pod "execpod-affinityr8q9l": Phase="Running", Reason="", readiness=true. Elapsed: 4.031516055s + Feb 25 11:13:06.541: INFO: Pod "execpod-affinityr8q9l" satisfied condition "running" + Feb 25 11:13:07.551: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9693 exec execpod-affinityr8q9l -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-nodeport-transition 80' + Feb 25 11:13:08.178: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-nodeport-transition 80\nConnection to affinity-nodeport-transition 80 port [tcp/http] succeeded!\n" + Feb 25 11:13:08.178: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:13:08.181: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9693 exec execpod-affinityr8q9l -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.37.115 80' + Feb 25 11:13:08.535: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.37.115 80\nConnection to 10.233.37.115 80 port [tcp/http] succeeded!\n" + Feb 25 11:13:08.535: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:13:08.535: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9693 exec execpod-affinityr8q9l -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.250 30798' + Feb 25 11:13:08.834: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.250 30798\nConnection to 192.168.121.250 30798 port [tcp/*] succeeded!\n" + Feb 25 11:13:08.834: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:13:08.835: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9693 exec execpod-affinityr8q9l -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.44 30798' + Feb 25 11:13:09.177: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.44 30798\nConnection to 192.168.121.44 30798 port [tcp/*] succeeded!\n" + Feb 25 11:13:09.177: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:13:09.318: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9693 exec execpod-affinityr8q9l -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.250:30798/ ; done' + Feb 25 11:13:09.853: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n" + Feb 25 11:13:09.854: INFO: stdout: "\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-pqxxm\naffinity-nodeport-transition-pqxxm\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-zk9lq\naffinity-nodeport-transition-zk9lq\naffinity-nodeport-transition-pqxxm\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-pqxxm\naffinity-nodeport-transition-pqxxm\naffinity-nodeport-transition-zk9lq\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn" + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-pqxxm + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-pqxxm + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-zk9lq + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-zk9lq + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-pqxxm + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-pqxxm + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-pqxxm + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-zk9lq + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:09.854: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:09.870: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9693 exec execpod-affinityr8q9l -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.250:30798/ ; done' + Feb 25 11:13:10.782: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:30798/\n" + Feb 25 11:13:10.782: INFO: stdout: "\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn\naffinity-nodeport-transition-c8vbn" + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Received response from host: affinity-nodeport-transition-c8vbn + Feb 25 11:13:10.783: INFO: Cleaning up the exec pod + STEP: deleting ReplicationController affinity-nodeport-transition in namespace services-9693, will wait for the garbage collector to delete the pods 02/25/23 11:13:10.814 + Feb 25 11:13:10.897: INFO: Deleting ReplicationController affinity-nodeport-transition took: 20.717446ms + Feb 25 11:13:10.998: INFO: Terminating ReplicationController affinity-nodeport-transition pods took: 100.539817ms + [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 16:46:16.398: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-7807" for this suite. 12/14/22 16:46:16.405 + Feb 25 11:13:14.237: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-9693" for this suite. 02/25/23 11:13:14.254 + [AfterEach] [sig-network] Services + test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -[sig-storage] Projected secret - should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:66 -[BeforeEach] [sig-storage] Projected secret +SSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-cli] Kubectl client Kubectl cluster-info + should check if Kubernetes control plane services is included in cluster-info [Conformance] + test/e2e/kubectl/kubectl.go:1248 +[BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:46:16.419 -Dec 14 16:46:16.419: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 16:46:16.426 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:46:16.452 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:46:16.456 -[It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:66 -STEP: Creating projection with secret that has name projected-secret-test-2e4b0575-fabd-4f83-a944-0a8ccadb3bf5 12/14/22 16:46:16.459 -STEP: Creating a pod to test consume secrets 12/14/22 16:46:16.466 -Dec 14 16:46:16.475: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615" in namespace "projected-1850" to be "Succeeded or Failed" -Dec 14 16:46:16.481: INFO: Pod "pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615": Phase="Pending", Reason="", readiness=false. Elapsed: 5.493521ms -Dec 14 16:46:18.490: INFO: Pod "pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014807259s -Dec 14 16:46:20.490: INFO: Pod "pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014048138s -STEP: Saw pod success 12/14/22 16:46:20.49 -Dec 14 16:46:20.490: INFO: Pod "pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615" satisfied condition "Succeeded or Failed" -Dec 14 16:46:20.496: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615 container projected-secret-volume-test: -STEP: delete the pod 12/14/22 16:46:20.507 -Dec 14 16:46:20.527: INFO: Waiting for pod pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615 to disappear -Dec 14 16:46:20.532: INFO: Pod pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615 no longer exists -[AfterEach] [sig-storage] Projected secret +STEP: Creating a kubernetes client 02/25/23 11:13:14.273 +Feb 25 11:13:14.273: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 11:13:14.275 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:14.307 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:14.316 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 +[It] should check if Kubernetes control plane services is included in cluster-info [Conformance] + test/e2e/kubectl/kubectl.go:1248 +STEP: validating cluster-info 02/25/23 11:13:14.324 +Feb 25 11:13:14.324: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-9327 cluster-info' +Feb 25 11:13:14.545: INFO: stderr: "" +Feb 25 11:13:14.545: INFO: stdout: "\x1b[0;32mKubernetes control plane\x1b[0m is running at \x1b[0;33mhttps://10.233.0.1:443\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n" +[AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 -Dec 14 16:46:20.533: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-1850" for this suite. 12/14/22 16:46:20.539 -{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance]","completed":246,"skipped":4546,"failed":0} +Feb 25 11:13:14.545: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-9327" for this suite. 02/25/23 11:13:14.553 +{"msg":"PASSED [sig-cli] Kubectl client Kubectl cluster-info should check if Kubernetes control plane services is included in cluster-info [Conformance]","completed":242,"skipped":4477,"failed":0} ------------------------------ -• [4.147 seconds] -[sig-storage] Projected secret -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:66 +• [0.301 seconds] +[sig-cli] Kubectl client +test/e2e/kubectl/framework.go:23 + Kubectl cluster-info + test/e2e/kubectl/kubectl.go:1242 + should check if Kubernetes control plane services is included in cluster-info [Conformance] + test/e2e/kubectl/kubectl.go:1248 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected secret - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:46:16.419 - Dec 14 16:46:16.419: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 16:46:16.426 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:46:16.452 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:46:16.456 - [It] should be consumable from pods in volume as non-root with defaultMode and fsGroup set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:66 - STEP: Creating projection with secret that has name projected-secret-test-2e4b0575-fabd-4f83-a944-0a8ccadb3bf5 12/14/22 16:46:16.459 - STEP: Creating a pod to test consume secrets 12/14/22 16:46:16.466 - Dec 14 16:46:16.475: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615" in namespace "projected-1850" to be "Succeeded or Failed" - Dec 14 16:46:16.481: INFO: Pod "pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615": Phase="Pending", Reason="", readiness=false. Elapsed: 5.493521ms - Dec 14 16:46:18.490: INFO: Pod "pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014807259s - Dec 14 16:46:20.490: INFO: Pod "pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014048138s - STEP: Saw pod success 12/14/22 16:46:20.49 - Dec 14 16:46:20.490: INFO: Pod "pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615" satisfied condition "Succeeded or Failed" - Dec 14 16:46:20.496: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615 container projected-secret-volume-test: - STEP: delete the pod 12/14/22 16:46:20.507 - Dec 14 16:46:20.527: INFO: Waiting for pod pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615 to disappear - Dec 14 16:46:20.532: INFO: Pod pod-projected-secrets-5cd51bfa-5c58-4b85-acec-dc8362fcc615 no longer exists - [AfterEach] [sig-storage] Projected secret + [BeforeEach] [sig-cli] Kubectl client + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 11:13:14.273 + Feb 25 11:13:14.273: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 11:13:14.275 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:14.307 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:14.316 + [BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 + [It] should check if Kubernetes control plane services is included in cluster-info [Conformance] + test/e2e/kubectl/kubectl.go:1248 + STEP: validating cluster-info 02/25/23 11:13:14.324 + Feb 25 11:13:14.324: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-9327 cluster-info' + Feb 25 11:13:14.545: INFO: stderr: "" + Feb 25 11:13:14.545: INFO: stdout: "\x1b[0;32mKubernetes control plane\x1b[0m is running at \x1b[0;33mhttps://10.233.0.1:443\x1b[0m\n\nTo further debug and diagnose cluster problems, use 'kubectl cluster-info dump'.\n" + [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 16:46:20.533: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-1850" for this suite. 12/14/22 16:46:20.539 + Feb 25 11:13:14.545: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-9327" for this suite. 02/25/23 11:13:14.553 << End Captured GinkgoWriter Output ------------------------------ -S +SSSSSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:196 -[BeforeEach] [sig-storage] EmptyDir volumes +[sig-storage] Secrets + should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:98 +[BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:46:20.568 -Dec 14 16:46:20.568: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir 12/14/22 16:46:20.571 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:46:20.591 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:46:20.595 -[It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:196 -STEP: Creating a pod to test emptydir 0644 on node default medium 12/14/22 16:46:20.6 -Dec 14 16:46:20.614: INFO: Waiting up to 5m0s for pod "pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5" in namespace "emptydir-3851" to be "Succeeded or Failed" -Dec 14 16:46:20.620: INFO: Pod "pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5": Phase="Pending", Reason="", readiness=false. Elapsed: 6.248077ms -Dec 14 16:46:22.627: INFO: Pod "pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013074084s -Dec 14 16:46:24.630: INFO: Pod "pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016179513s -STEP: Saw pod success 12/14/22 16:46:24.63 -Dec 14 16:46:24.632: INFO: Pod "pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5" satisfied condition "Succeeded or Failed" -Dec 14 16:46:24.639: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5 container test-container: -STEP: delete the pod 12/14/22 16:46:24.649 -Dec 14 16:46:24.670: INFO: Waiting for pod pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5 to disappear -Dec 14 16:46:24.676: INFO: Pod pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5 no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +STEP: Creating a kubernetes client 02/25/23 11:13:14.575 +Feb 25 11:13:14.575: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename secrets 02/25/23 11:13:14.578 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:14.622 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:14.639 +[It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:98 +STEP: Creating secret with name secret-test-d47912cd-8be4-47f4-9718-7943bfecde0e 02/25/23 11:13:14.713 +STEP: Creating a pod to test consume secrets 02/25/23 11:13:14.724 +Feb 25 11:13:14.746: INFO: Waiting up to 5m0s for pod "pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b" in namespace "secrets-7114" to be "Succeeded or Failed" +Feb 25 11:13:14.753: INFO: Pod "pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b": Phase="Pending", Reason="", readiness=false. Elapsed: 6.927429ms +Feb 25 11:13:16.761: INFO: Pod "pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015514547s +Feb 25 11:13:18.761: INFO: Pod "pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.015439714s +Feb 25 11:13:20.762: INFO: Pod "pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016078228s +STEP: Saw pod success 02/25/23 11:13:20.762 +Feb 25 11:13:20.762: INFO: Pod "pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b" satisfied condition "Succeeded or Failed" +Feb 25 11:13:20.768: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b container secret-volume-test: +STEP: delete the pod 02/25/23 11:13:20.783 +Feb 25 11:13:20.808: INFO: Waiting for pod pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b to disappear +Feb 25 11:13:20.815: INFO: Pod pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b no longer exists +[AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 -Dec 14 16:46:24.676: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-3851" for this suite. 12/14/22 16:46:24.684 -{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","completed":247,"skipped":4547,"failed":0} +Feb 25 11:13:20.815: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-7114" for this suite. 02/25/23 11:13:20.837 +STEP: Destroying namespace "secret-namespace-9390" for this suite. 02/25/23 11:13:20.858 +{"msg":"PASSED [sig-storage] Secrets should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance]","completed":243,"skipped":4487,"failed":0} ------------------------------ -• [4.125 seconds] -[sig-storage] EmptyDir volumes +• [SLOW TEST] [6.331 seconds] +[sig-storage] Secrets test/e2e/common/storage/framework.go:23 - should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:196 + should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:98 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir volumes + [BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:46:20.568 - Dec 14 16:46:20.568: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir 12/14/22 16:46:20.571 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:46:20.591 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:46:20.595 - [It] should support (non-root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:196 - STEP: Creating a pod to test emptydir 0644 on node default medium 12/14/22 16:46:20.6 - Dec 14 16:46:20.614: INFO: Waiting up to 5m0s for pod "pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5" in namespace "emptydir-3851" to be "Succeeded or Failed" - Dec 14 16:46:20.620: INFO: Pod "pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5": Phase="Pending", Reason="", readiness=false. Elapsed: 6.248077ms - Dec 14 16:46:22.627: INFO: Pod "pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013074084s - Dec 14 16:46:24.630: INFO: Pod "pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016179513s - STEP: Saw pod success 12/14/22 16:46:24.63 - Dec 14 16:46:24.632: INFO: Pod "pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5" satisfied condition "Succeeded or Failed" - Dec 14 16:46:24.639: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5 container test-container: - STEP: delete the pod 12/14/22 16:46:24.649 - Dec 14 16:46:24.670: INFO: Waiting for pod pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5 to disappear - Dec 14 16:46:24.676: INFO: Pod pod-bab6ed04-c027-4a79-8e0b-ea7389a4c7b5 no longer exists - [AfterEach] [sig-storage] EmptyDir volumes + STEP: Creating a kubernetes client 02/25/23 11:13:14.575 + Feb 25 11:13:14.575: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename secrets 02/25/23 11:13:14.578 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:14.622 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:14.639 + [It] should be able to mount in a volume regardless of a different secret existing with same name in different namespace [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:98 + STEP: Creating secret with name secret-test-d47912cd-8be4-47f4-9718-7943bfecde0e 02/25/23 11:13:14.713 + STEP: Creating a pod to test consume secrets 02/25/23 11:13:14.724 + Feb 25 11:13:14.746: INFO: Waiting up to 5m0s for pod "pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b" in namespace "secrets-7114" to be "Succeeded or Failed" + Feb 25 11:13:14.753: INFO: Pod "pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b": Phase="Pending", Reason="", readiness=false. Elapsed: 6.927429ms + Feb 25 11:13:16.761: INFO: Pod "pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015514547s + Feb 25 11:13:18.761: INFO: Pod "pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.015439714s + Feb 25 11:13:20.762: INFO: Pod "pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016078228s + STEP: Saw pod success 02/25/23 11:13:20.762 + Feb 25 11:13:20.762: INFO: Pod "pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b" satisfied condition "Succeeded or Failed" + Feb 25 11:13:20.768: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b container secret-volume-test: + STEP: delete the pod 02/25/23 11:13:20.783 + Feb 25 11:13:20.808: INFO: Waiting for pod pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b to disappear + Feb 25 11:13:20.815: INFO: Pod pod-secrets-ac10e3ec-c740-4f86-8cc4-a2973e51609b no longer exists + [AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 - Dec 14 16:46:24.676: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-3851" for this suite. 12/14/22 16:46:24.684 + Feb 25 11:13:20.815: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "secrets-7114" for this suite. 02/25/23 11:13:20.837 + STEP: Destroying namespace "secret-namespace-9390" for this suite. 02/25/23 11:13:20.858 << End Captured GinkgoWriter Output ------------------------------ -S +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Probing container - should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:131 -[BeforeEach] [sig-node] Probing container +[sig-node] Sysctls [LinuxOnly] [NodeConformance] + should support sysctls [MinimumKubeletVersion:1.21] [Conformance] + test/e2e/common/node/sysctl.go:77 +[BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] + test/e2e/common/node/sysctl.go:37 +[BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:46:24.699 -Dec 14 16:46:24.700: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-probe 12/14/22 16:46:24.705 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:46:24.731 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:46:24.739 -[BeforeEach] [sig-node] Probing container - test/e2e/common/node/container_probe.go:59 -[It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:131 -STEP: Creating pod busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375 in namespace container-probe-62 12/14/22 16:46:24.745 -Dec 14 16:46:24.756: INFO: Waiting up to 5m0s for pod "busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375" in namespace "container-probe-62" to be "not pending" -Dec 14 16:46:24.768: INFO: Pod "busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375": Phase="Pending", Reason="", readiness=false. Elapsed: 11.345072ms -Dec 14 16:46:26.777: INFO: Pod "busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375": Phase="Running", Reason="", readiness=true. Elapsed: 2.020513369s -Dec 14 16:46:26.777: INFO: Pod "busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375" satisfied condition "not pending" -Dec 14 16:46:26.777: INFO: Started pod busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375 in namespace container-probe-62 -STEP: checking the pod's current state and verifying that restartCount is present 12/14/22 16:46:26.777 -Dec 14 16:46:26.782: INFO: Initial restart count of pod busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375 is 0 -Dec 14 16:47:17.003: INFO: Restart count of pod container-probe-62/busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375 is now 1 (50.22068063s elapsed) -STEP: deleting the pod 12/14/22 16:47:17.004 -[AfterEach] [sig-node] Probing container +STEP: Creating a kubernetes client 02/25/23 11:13:20.917 +Feb 25 11:13:20.917: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename sysctl 02/25/23 11:13:20.922 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:20.965 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:20.969 +[BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] + test/e2e/common/node/sysctl.go:67 +[It] should support sysctls [MinimumKubeletVersion:1.21] [Conformance] + test/e2e/common/node/sysctl.go:77 +STEP: Creating a pod with the kernel.shm_rmid_forced sysctl 02/25/23 11:13:20.973 +STEP: Watching for error events or started pod 02/25/23 11:13:20.995 +STEP: Waiting for pod completion 02/25/23 11:13:23.003 +Feb 25 11:13:23.003: INFO: Waiting up to 3m0s for pod "sysctl-a6003562-b848-4783-a526-d3769ea7319c" in namespace "sysctl-6994" to be "completed" +Feb 25 11:13:23.008: INFO: Pod "sysctl-a6003562-b848-4783-a526-d3769ea7319c": Phase="Pending", Reason="", readiness=false. Elapsed: 5.369912ms +Feb 25 11:13:25.020: INFO: Pod "sysctl-a6003562-b848-4783-a526-d3769ea7319c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016618063s +Feb 25 11:13:27.017: INFO: Pod "sysctl-a6003562-b848-4783-a526-d3769ea7319c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.0136503s +Feb 25 11:13:27.017: INFO: Pod "sysctl-a6003562-b848-4783-a526-d3769ea7319c" satisfied condition "completed" +STEP: Checking that the pod succeeded 02/25/23 11:13:27.022 +STEP: Getting logs from the pod 02/25/23 11:13:27.022 +STEP: Checking that the sysctl is actually updated 02/25/23 11:13:27.032 +[AfterEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] test/e2e/framework/framework.go:187 -Dec 14 16:47:17.030: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-probe-62" for this suite. 12/14/22 16:47:17.037 -{"msg":"PASSED [sig-node] Probing container should be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","completed":248,"skipped":4548,"failed":0} +Feb 25 11:13:27.033: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "sysctl-6994" for this suite. 02/25/23 11:13:27.04 +{"msg":"PASSED [sig-node] Sysctls [LinuxOnly] [NodeConformance] should support sysctls [MinimumKubeletVersion:1.21] [Conformance]","completed":244,"skipped":4551,"failed":0} ------------------------------ -• [SLOW TEST] [52.347 seconds] -[sig-node] Probing container +• [SLOW TEST] [6.134 seconds] +[sig-node] Sysctls [LinuxOnly] [NodeConformance] test/e2e/common/node/framework.go:23 - should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:131 + should support sysctls [MinimumKubeletVersion:1.21] [Conformance] + test/e2e/common/node/sysctl.go:77 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Probing container + [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] + test/e2e/common/node/sysctl.go:37 + [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:46:24.699 - Dec 14 16:46:24.700: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-probe 12/14/22 16:46:24.705 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:46:24.731 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:46:24.739 - [BeforeEach] [sig-node] Probing container - test/e2e/common/node/container_probe.go:59 - [It] should be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:131 - STEP: Creating pod busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375 in namespace container-probe-62 12/14/22 16:46:24.745 - Dec 14 16:46:24.756: INFO: Waiting up to 5m0s for pod "busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375" in namespace "container-probe-62" to be "not pending" - Dec 14 16:46:24.768: INFO: Pod "busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375": Phase="Pending", Reason="", readiness=false. Elapsed: 11.345072ms - Dec 14 16:46:26.777: INFO: Pod "busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375": Phase="Running", Reason="", readiness=true. Elapsed: 2.020513369s - Dec 14 16:46:26.777: INFO: Pod "busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375" satisfied condition "not pending" - Dec 14 16:46:26.777: INFO: Started pod busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375 in namespace container-probe-62 - STEP: checking the pod's current state and verifying that restartCount is present 12/14/22 16:46:26.777 - Dec 14 16:46:26.782: INFO: Initial restart count of pod busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375 is 0 - Dec 14 16:47:17.003: INFO: Restart count of pod container-probe-62/busybox-a7c4e881-3318-4da0-8bf2-b60b5f3e6375 is now 1 (50.22068063s elapsed) - STEP: deleting the pod 12/14/22 16:47:17.004 - [AfterEach] [sig-node] Probing container + STEP: Creating a kubernetes client 02/25/23 11:13:20.917 + Feb 25 11:13:20.917: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename sysctl 02/25/23 11:13:20.922 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:20.965 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:20.969 + [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] + test/e2e/common/node/sysctl.go:67 + [It] should support sysctls [MinimumKubeletVersion:1.21] [Conformance] + test/e2e/common/node/sysctl.go:77 + STEP: Creating a pod with the kernel.shm_rmid_forced sysctl 02/25/23 11:13:20.973 + STEP: Watching for error events or started pod 02/25/23 11:13:20.995 + STEP: Waiting for pod completion 02/25/23 11:13:23.003 + Feb 25 11:13:23.003: INFO: Waiting up to 3m0s for pod "sysctl-a6003562-b848-4783-a526-d3769ea7319c" in namespace "sysctl-6994" to be "completed" + Feb 25 11:13:23.008: INFO: Pod "sysctl-a6003562-b848-4783-a526-d3769ea7319c": Phase="Pending", Reason="", readiness=false. Elapsed: 5.369912ms + Feb 25 11:13:25.020: INFO: Pod "sysctl-a6003562-b848-4783-a526-d3769ea7319c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016618063s + Feb 25 11:13:27.017: INFO: Pod "sysctl-a6003562-b848-4783-a526-d3769ea7319c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.0136503s + Feb 25 11:13:27.017: INFO: Pod "sysctl-a6003562-b848-4783-a526-d3769ea7319c" satisfied condition "completed" + STEP: Checking that the pod succeeded 02/25/23 11:13:27.022 + STEP: Getting logs from the pod 02/25/23 11:13:27.022 + STEP: Checking that the sysctl is actually updated 02/25/23 11:13:27.032 + [AfterEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] test/e2e/framework/framework.go:187 - Dec 14 16:47:17.030: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-probe-62" for this suite. 12/14/22 16:47:17.037 + Feb 25 11:13:27.033: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "sysctl-6994" for this suite. 02/25/23 11:13:27.04 << End Captured GinkgoWriter Output ------------------------------ -[sig-node] Container Runtime blackbox test on terminated container - should report termination message if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:194 -[BeforeEach] [sig-node] Container Runtime +SSSSSSSSSSSSSS +------------------------------ +[sig-node] Containers + should use the image defaults if command and args are blank [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:38 +[BeforeEach] [sig-node] Containers test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:47:17.048 -Dec 14 16:47:17.052: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-runtime 12/14/22 16:47:17.066 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:47:17.111 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:47:17.113 -[It] should report termination message if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:194 -STEP: create the container 12/14/22 16:47:17.117 -STEP: wait for the container to reach Succeeded 12/14/22 16:47:17.132 -STEP: get the container status 12/14/22 16:47:20.184 -STEP: the container should be terminated 12/14/22 16:47:20.188 -STEP: the termination message should be set 12/14/22 16:47:20.189 -Dec 14 16:47:20.189: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- -STEP: delete the container 12/14/22 16:47:20.189 -[AfterEach] [sig-node] Container Runtime +STEP: Creating a kubernetes client 02/25/23 11:13:27.053 +Feb 25 11:13:27.053: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename containers 02/25/23 11:13:27.056 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:27.077 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:27.082 +[It] should use the image defaults if command and args are blank [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:38 +Feb 25 11:13:27.103: INFO: Waiting up to 5m0s for pod "client-containers-963b1251-845a-45ef-a908-77b2d14f7558" in namespace "containers-5135" to be "running" +Feb 25 11:13:27.109: INFO: Pod "client-containers-963b1251-845a-45ef-a908-77b2d14f7558": Phase="Pending", Reason="", readiness=false. Elapsed: 6.100694ms +Feb 25 11:13:29.117: INFO: Pod "client-containers-963b1251-845a-45ef-a908-77b2d14f7558": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014166004s +Feb 25 11:13:31.121: INFO: Pod "client-containers-963b1251-845a-45ef-a908-77b2d14f7558": Phase="Running", Reason="", readiness=true. Elapsed: 4.017891269s +Feb 25 11:13:31.121: INFO: Pod "client-containers-963b1251-845a-45ef-a908-77b2d14f7558" satisfied condition "running" +[AfterEach] [sig-node] Containers test/e2e/framework/framework.go:187 -Dec 14 16:47:20.220: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-runtime-1251" for this suite. 12/14/22 16:47:20.235 -{"msg":"PASSED [sig-node] Container Runtime blackbox test on terminated container should report termination message if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance]","completed":249,"skipped":4548,"failed":0} +Feb 25 11:13:31.140: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "containers-5135" for this suite. 02/25/23 11:13:31.153 +{"msg":"PASSED [sig-node] Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance]","completed":245,"skipped":4565,"failed":0} ------------------------------ -• [3.203 seconds] -[sig-node] Container Runtime +• [4.123 seconds] +[sig-node] Containers test/e2e/common/node/framework.go:23 - blackbox test - test/e2e/common/node/runtime.go:43 - on terminated container - test/e2e/common/node/runtime.go:136 - should report termination message if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:194 + should use the image defaults if command and args are blank [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:38 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Container Runtime + [BeforeEach] [sig-node] Containers test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:47:17.048 - Dec 14 16:47:17.052: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-runtime 12/14/22 16:47:17.066 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:47:17.111 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:47:17.113 - [It] should report termination message if TerminationMessagePath is set as non-root user and at a non-default path [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:194 - STEP: create the container 12/14/22 16:47:17.117 - STEP: wait for the container to reach Succeeded 12/14/22 16:47:17.132 - STEP: get the container status 12/14/22 16:47:20.184 - STEP: the container should be terminated 12/14/22 16:47:20.188 - STEP: the termination message should be set 12/14/22 16:47:20.189 - Dec 14 16:47:20.189: INFO: Expected: &{DONE} to match Container's Termination Message: DONE -- - STEP: delete the container 12/14/22 16:47:20.189 - [AfterEach] [sig-node] Container Runtime + STEP: Creating a kubernetes client 02/25/23 11:13:27.053 + Feb 25 11:13:27.053: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename containers 02/25/23 11:13:27.056 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:27.077 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:27.082 + [It] should use the image defaults if command and args are blank [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:38 + Feb 25 11:13:27.103: INFO: Waiting up to 5m0s for pod "client-containers-963b1251-845a-45ef-a908-77b2d14f7558" in namespace "containers-5135" to be "running" + Feb 25 11:13:27.109: INFO: Pod "client-containers-963b1251-845a-45ef-a908-77b2d14f7558": Phase="Pending", Reason="", readiness=false. Elapsed: 6.100694ms + Feb 25 11:13:29.117: INFO: Pod "client-containers-963b1251-845a-45ef-a908-77b2d14f7558": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014166004s + Feb 25 11:13:31.121: INFO: Pod "client-containers-963b1251-845a-45ef-a908-77b2d14f7558": Phase="Running", Reason="", readiness=true. Elapsed: 4.017891269s + Feb 25 11:13:31.121: INFO: Pod "client-containers-963b1251-845a-45ef-a908-77b2d14f7558" satisfied condition "running" + [AfterEach] [sig-node] Containers test/e2e/framework/framework.go:187 - Dec 14 16:47:20.220: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-runtime-1251" for this suite. 12/14/22 16:47:20.235 + Feb 25 11:13:31.140: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "containers-5135" for this suite. 02/25/23 11:13:31.153 << End Captured GinkgoWriter Output ------------------------------ -SS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-network] Services - should be able to change the type from ExternalName to ClusterIP [Conformance] - test/e2e/network/service.go:1404 -[BeforeEach] [sig-network] Services +[sig-network] Proxy version v1 + A set of valid responses are returned for both pod and service Proxy [Conformance] + test/e2e/network/proxy.go:380 +[BeforeEach] version v1 test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:47:20.252 -Dec 14 16:47:20.252: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 16:47:20.254 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:47:20.282 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:47:20.286 -[BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 -[It] should be able to change the type from ExternalName to ClusterIP [Conformance] - test/e2e/network/service.go:1404 -STEP: creating a service externalname-service with the type=ExternalName in namespace services-3918 12/14/22 16:47:20.289 -STEP: changing the ExternalName service to type=ClusterIP 12/14/22 16:47:20.299 -STEP: creating replication controller externalname-service in namespace services-3918 12/14/22 16:47:20.32 -I1214 16:47:20.330465 14 runners.go:193] Created replication controller with name: externalname-service, namespace: services-3918, replica count: 2 -I1214 16:47:23.381646 14 runners.go:193] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -Dec 14 16:47:23.381: INFO: Creating new exec pod -Dec 14 16:47:23.398: INFO: Waiting up to 5m0s for pod "execpodwdjsm" in namespace "services-3918" to be "running" -Dec 14 16:47:23.406: INFO: Pod "execpodwdjsm": Phase="Pending", Reason="", readiness=false. Elapsed: 7.649548ms -Dec 14 16:47:25.415: INFO: Pod "execpodwdjsm": Phase="Running", Reason="", readiness=true. Elapsed: 2.016561101s -Dec 14 16:47:25.415: INFO: Pod "execpodwdjsm" satisfied condition "running" -Dec 14 16:47:26.416: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3918 exec execpodwdjsm -- /bin/sh -x -c echo hostName | nc -v -t -w 2 externalname-service 80' -Dec 14 16:47:26.745: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n" -Dec 14 16:47:26.745: INFO: stdout: "externalname-service-9nsx7" -Dec 14 16:47:26.745: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3918 exec execpodwdjsm -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.37.99 80' -Dec 14 16:47:27.047: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.37.99 80\nConnection to 10.233.37.99 80 port [tcp/http] succeeded!\n" -Dec 14 16:47:27.047: INFO: stdout: "externalname-service-9nsx7" -Dec 14 16:47:27.047: INFO: Cleaning up the ExternalName to ClusterIP test service -[AfterEach] [sig-network] Services +STEP: Creating a kubernetes client 02/25/23 11:13:31.182 +Feb 25 11:13:31.182: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename proxy 02/25/23 11:13:31.185 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:31.228 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:31.233 +[It] A set of valid responses are returned for both pod and service Proxy [Conformance] + test/e2e/network/proxy.go:380 +Feb 25 11:13:31.238: INFO: Creating pod... +Feb 25 11:13:31.260: INFO: Waiting up to 5m0s for pod "agnhost" in namespace "proxy-5374" to be "running" +Feb 25 11:13:31.268: INFO: Pod "agnhost": Phase="Pending", Reason="", readiness=false. Elapsed: 8.055986ms +Feb 25 11:13:33.287: INFO: Pod "agnhost": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026977248s +Feb 25 11:13:35.278: INFO: Pod "agnhost": Phase="Running", Reason="", readiness=true. Elapsed: 4.018341225s +Feb 25 11:13:35.278: INFO: Pod "agnhost" satisfied condition "running" +Feb 25 11:13:35.278: INFO: Creating service... +Feb 25 11:13:35.298: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/pods/agnhost/proxy?method=DELETE +Feb 25 11:13:35.314: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE +Feb 25 11:13:35.314: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/pods/agnhost/proxy?method=OPTIONS +Feb 25 11:13:35.324: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS +Feb 25 11:13:35.324: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/pods/agnhost/proxy?method=PATCH +Feb 25 11:13:35.333: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH +Feb 25 11:13:35.333: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/pods/agnhost/proxy?method=POST +Feb 25 11:13:35.342: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST +Feb 25 11:13:35.342: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/pods/agnhost/proxy?method=PUT +Feb 25 11:13:35.352: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT +Feb 25 11:13:35.352: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/services/e2e-proxy-test-service/proxy?method=DELETE +Feb 25 11:13:35.375: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE +Feb 25 11:13:35.375: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/services/e2e-proxy-test-service/proxy?method=OPTIONS +Feb 25 11:13:35.386: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS +Feb 25 11:13:35.386: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/services/e2e-proxy-test-service/proxy?method=PATCH +Feb 25 11:13:35.403: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH +Feb 25 11:13:35.404: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/services/e2e-proxy-test-service/proxy?method=POST +Feb 25 11:13:35.418: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST +Feb 25 11:13:35.418: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/services/e2e-proxy-test-service/proxy?method=PUT +Feb 25 11:13:35.433: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT +Feb 25 11:13:35.433: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/pods/agnhost/proxy?method=GET +Feb 25 11:13:35.440: INFO: http.Client request:GET StatusCode:301 +Feb 25 11:13:35.440: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/services/e2e-proxy-test-service/proxy?method=GET +Feb 25 11:13:35.463: INFO: http.Client request:GET StatusCode:301 +Feb 25 11:13:35.463: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/pods/agnhost/proxy?method=HEAD +Feb 25 11:13:35.474: INFO: http.Client request:HEAD StatusCode:301 +Feb 25 11:13:35.474: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/services/e2e-proxy-test-service/proxy?method=HEAD +Feb 25 11:13:35.487: INFO: http.Client request:HEAD StatusCode:301 +[AfterEach] version v1 test/e2e/framework/framework.go:187 -Dec 14 16:47:27.073: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-3918" for this suite. 12/14/22 16:47:27.095 -[AfterEach] [sig-network] Services - test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should be able to change the type from ExternalName to ClusterIP [Conformance]","completed":250,"skipped":4550,"failed":0} +Feb 25 11:13:35.487: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "proxy-5374" for this suite. 02/25/23 11:13:35.498 +{"msg":"PASSED [sig-network] Proxy version v1 A set of valid responses are returned for both pod and service Proxy [Conformance]","completed":246,"skipped":4655,"failed":0} ------------------------------ -• [SLOW TEST] [6.853 seconds] -[sig-network] Services +• [4.341 seconds] +[sig-network] Proxy test/e2e/network/common/framework.go:23 - should be able to change the type from ExternalName to ClusterIP [Conformance] - test/e2e/network/service.go:1404 + version v1 + test/e2e/network/proxy.go:74 + A set of valid responses are returned for both pod and service Proxy [Conformance] + test/e2e/network/proxy.go:380 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Services + [BeforeEach] version v1 test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:47:20.252 - Dec 14 16:47:20.252: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 16:47:20.254 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:47:20.282 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:47:20.286 - [BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 - [It] should be able to change the type from ExternalName to ClusterIP [Conformance] - test/e2e/network/service.go:1404 - STEP: creating a service externalname-service with the type=ExternalName in namespace services-3918 12/14/22 16:47:20.289 - STEP: changing the ExternalName service to type=ClusterIP 12/14/22 16:47:20.299 - STEP: creating replication controller externalname-service in namespace services-3918 12/14/22 16:47:20.32 - I1214 16:47:20.330465 14 runners.go:193] Created replication controller with name: externalname-service, namespace: services-3918, replica count: 2 - I1214 16:47:23.381646 14 runners.go:193] externalname-service Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - Dec 14 16:47:23.381: INFO: Creating new exec pod - Dec 14 16:47:23.398: INFO: Waiting up to 5m0s for pod "execpodwdjsm" in namespace "services-3918" to be "running" - Dec 14 16:47:23.406: INFO: Pod "execpodwdjsm": Phase="Pending", Reason="", readiness=false. Elapsed: 7.649548ms - Dec 14 16:47:25.415: INFO: Pod "execpodwdjsm": Phase="Running", Reason="", readiness=true. Elapsed: 2.016561101s - Dec 14 16:47:25.415: INFO: Pod "execpodwdjsm" satisfied condition "running" - Dec 14 16:47:26.416: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3918 exec execpodwdjsm -- /bin/sh -x -c echo hostName | nc -v -t -w 2 externalname-service 80' - Dec 14 16:47:26.745: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 externalname-service 80\nConnection to externalname-service 80 port [tcp/http] succeeded!\n" - Dec 14 16:47:26.745: INFO: stdout: "externalname-service-9nsx7" - Dec 14 16:47:26.745: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-3918 exec execpodwdjsm -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.37.99 80' - Dec 14 16:47:27.047: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.37.99 80\nConnection to 10.233.37.99 80 port [tcp/http] succeeded!\n" - Dec 14 16:47:27.047: INFO: stdout: "externalname-service-9nsx7" - Dec 14 16:47:27.047: INFO: Cleaning up the ExternalName to ClusterIP test service - [AfterEach] [sig-network] Services + STEP: Creating a kubernetes client 02/25/23 11:13:31.182 + Feb 25 11:13:31.182: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename proxy 02/25/23 11:13:31.185 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:31.228 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:31.233 + [It] A set of valid responses are returned for both pod and service Proxy [Conformance] + test/e2e/network/proxy.go:380 + Feb 25 11:13:31.238: INFO: Creating pod... + Feb 25 11:13:31.260: INFO: Waiting up to 5m0s for pod "agnhost" in namespace "proxy-5374" to be "running" + Feb 25 11:13:31.268: INFO: Pod "agnhost": Phase="Pending", Reason="", readiness=false. Elapsed: 8.055986ms + Feb 25 11:13:33.287: INFO: Pod "agnhost": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026977248s + Feb 25 11:13:35.278: INFO: Pod "agnhost": Phase="Running", Reason="", readiness=true. Elapsed: 4.018341225s + Feb 25 11:13:35.278: INFO: Pod "agnhost" satisfied condition "running" + Feb 25 11:13:35.278: INFO: Creating service... + Feb 25 11:13:35.298: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/pods/agnhost/proxy?method=DELETE + Feb 25 11:13:35.314: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE + Feb 25 11:13:35.314: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/pods/agnhost/proxy?method=OPTIONS + Feb 25 11:13:35.324: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS + Feb 25 11:13:35.324: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/pods/agnhost/proxy?method=PATCH + Feb 25 11:13:35.333: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH + Feb 25 11:13:35.333: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/pods/agnhost/proxy?method=POST + Feb 25 11:13:35.342: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST + Feb 25 11:13:35.342: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/pods/agnhost/proxy?method=PUT + Feb 25 11:13:35.352: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT + Feb 25 11:13:35.352: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/services/e2e-proxy-test-service/proxy?method=DELETE + Feb 25 11:13:35.375: INFO: http.Client request:DELETE | StatusCode:200 | Response:foo | Method:DELETE + Feb 25 11:13:35.375: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/services/e2e-proxy-test-service/proxy?method=OPTIONS + Feb 25 11:13:35.386: INFO: http.Client request:OPTIONS | StatusCode:200 | Response:foo | Method:OPTIONS + Feb 25 11:13:35.386: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/services/e2e-proxy-test-service/proxy?method=PATCH + Feb 25 11:13:35.403: INFO: http.Client request:PATCH | StatusCode:200 | Response:foo | Method:PATCH + Feb 25 11:13:35.404: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/services/e2e-proxy-test-service/proxy?method=POST + Feb 25 11:13:35.418: INFO: http.Client request:POST | StatusCode:200 | Response:foo | Method:POST + Feb 25 11:13:35.418: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/services/e2e-proxy-test-service/proxy?method=PUT + Feb 25 11:13:35.433: INFO: http.Client request:PUT | StatusCode:200 | Response:foo | Method:PUT + Feb 25 11:13:35.433: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/pods/agnhost/proxy?method=GET + Feb 25 11:13:35.440: INFO: http.Client request:GET StatusCode:301 + Feb 25 11:13:35.440: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/services/e2e-proxy-test-service/proxy?method=GET + Feb 25 11:13:35.463: INFO: http.Client request:GET StatusCode:301 + Feb 25 11:13:35.463: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/pods/agnhost/proxy?method=HEAD + Feb 25 11:13:35.474: INFO: http.Client request:HEAD StatusCode:301 + Feb 25 11:13:35.474: INFO: Starting http.Client for https://10.233.0.1:443/api/v1/namespaces/proxy-5374/services/e2e-proxy-test-service/proxy?method=HEAD + Feb 25 11:13:35.487: INFO: http.Client request:HEAD StatusCode:301 + [AfterEach] version v1 test/e2e/framework/framework.go:187 - Dec 14 16:47:27.073: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-3918" for this suite. 12/14/22 16:47:27.095 - [AfterEach] [sig-network] Services - test/e2e/network/service.go:762 + Feb 25 11:13:35.487: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "proxy-5374" for this suite. 02/25/23 11:13:35.498 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +S ------------------------------ [sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [Conformance] test/e2e/storage/subpath.go:60 [BeforeEach] [sig-storage] Subpath test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:47:27.108 -Dec 14 16:47:27.108: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename subpath 12/14/22 16:47:27.117 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:47:27.154 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:47:27.157 +STEP: Creating a kubernetes client 02/25/23 11:13:35.525 +Feb 25 11:13:35.525: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename subpath 02/25/23 11:13:35.529 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:35.57 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:35.576 [BeforeEach] Atomic writer volumes test/e2e/storage/subpath.go:40 -STEP: Setting up data 12/14/22 16:47:27.161 +STEP: Setting up data 02/25/23 11:13:35.581 [It] should support subpaths with secret pod [Conformance] test/e2e/storage/subpath.go:60 -STEP: Creating pod pod-subpath-test-secret-n7f4 12/14/22 16:47:27.198 -STEP: Creating a pod to test atomic-volume-subpath 12/14/22 16:47:27.198 -Dec 14 16:47:27.217: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-n7f4" in namespace "subpath-3167" to be "Succeeded or Failed" -Dec 14 16:47:27.224: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Pending", Reason="", readiness=false. Elapsed: 6.723888ms -Dec 14 16:47:29.231: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 2.014069242s -Dec 14 16:47:31.230: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 4.013072841s -Dec 14 16:47:33.234: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 6.016520326s -Dec 14 16:47:35.231: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 8.013831662s -Dec 14 16:47:37.231: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 10.01370429s -Dec 14 16:47:39.233: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 12.015952509s -Dec 14 16:47:41.231: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 14.014035426s -Dec 14 16:47:43.230: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 16.013110031s -Dec 14 16:47:45.237: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 18.019907593s -Dec 14 16:47:47.231: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 20.013647413s -Dec 14 16:47:49.232: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=false. Elapsed: 22.01467481s -Dec 14 16:47:51.232: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.014581866s -STEP: Saw pod success 12/14/22 16:47:51.232 -Dec 14 16:47:51.232: INFO: Pod "pod-subpath-test-secret-n7f4" satisfied condition "Succeeded or Failed" -Dec 14 16:47:51.237: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-subpath-test-secret-n7f4 container test-container-subpath-secret-n7f4: -STEP: delete the pod 12/14/22 16:47:51.248 -Dec 14 16:47:51.267: INFO: Waiting for pod pod-subpath-test-secret-n7f4 to disappear -Dec 14 16:47:51.272: INFO: Pod pod-subpath-test-secret-n7f4 no longer exists -STEP: Deleting pod pod-subpath-test-secret-n7f4 12/14/22 16:47:51.272 -Dec 14 16:47:51.272: INFO: Deleting pod "pod-subpath-test-secret-n7f4" in namespace "subpath-3167" +STEP: Creating pod pod-subpath-test-secret-rsr2 02/25/23 11:13:35.61 +STEP: Creating a pod to test atomic-volume-subpath 02/25/23 11:13:35.61 +Feb 25 11:13:35.634: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-rsr2" in namespace "subpath-7797" to be "Succeeded or Failed" +Feb 25 11:13:35.649: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Pending", Reason="", readiness=false. Elapsed: 15.11444ms +Feb 25 11:13:37.658: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 2.024082612s +Feb 25 11:13:39.673: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 4.038289298s +Feb 25 11:13:41.658: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 6.023656371s +Feb 25 11:13:43.660: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 8.025167386s +Feb 25 11:13:45.671: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 10.036309485s +Feb 25 11:13:47.658: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 12.024031526s +Feb 25 11:13:49.665: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 14.030428174s +Feb 25 11:13:51.663: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 16.028908607s +Feb 25 11:13:53.657: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 18.022516341s +Feb 25 11:13:55.657: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 20.022614664s +Feb 25 11:13:57.658: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=false. Elapsed: 22.023371834s +Feb 25 11:13:59.659: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.024637256s +STEP: Saw pod success 02/25/23 11:13:59.659 +Feb 25 11:13:59.660: INFO: Pod "pod-subpath-test-secret-rsr2" satisfied condition "Succeeded or Failed" +Feb 25 11:13:59.666: INFO: Trying to get logs from node diek3dua9gay-2 pod pod-subpath-test-secret-rsr2 container test-container-subpath-secret-rsr2: +STEP: delete the pod 02/25/23 11:13:59.694 +Feb 25 11:13:59.716: INFO: Waiting for pod pod-subpath-test-secret-rsr2 to disappear +Feb 25 11:13:59.721: INFO: Pod pod-subpath-test-secret-rsr2 no longer exists +STEP: Deleting pod pod-subpath-test-secret-rsr2 02/25/23 11:13:59.721 +Feb 25 11:13:59.721: INFO: Deleting pod "pod-subpath-test-secret-rsr2" in namespace "subpath-7797" [AfterEach] [sig-storage] Subpath test/e2e/framework/framework.go:187 -Dec 14 16:47:51.276: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "subpath-3167" for this suite. 12/14/22 16:47:51.282 -{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [Conformance]","completed":251,"skipped":4600,"failed":0} +Feb 25 11:13:59.726: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "subpath-7797" for this suite. 02/25/23 11:13:59.733 +{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with secret pod [Conformance]","completed":247,"skipped":4656,"failed":0} ------------------------------ -• [SLOW TEST] [24.181 seconds] +• [SLOW TEST] [24.221 seconds] [sig-storage] Subpath test/e2e/storage/utils/framework.go:23 Atomic writer volumes @@ -22365,8302 +23059,9127 @@ test/e2e/storage/utils/framework.go:23 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-storage] Subpath test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:47:27.108 - Dec 14 16:47:27.108: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename subpath 12/14/22 16:47:27.117 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:47:27.154 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:47:27.157 + STEP: Creating a kubernetes client 02/25/23 11:13:35.525 + Feb 25 11:13:35.525: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename subpath 02/25/23 11:13:35.529 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:35.57 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:35.576 [BeforeEach] Atomic writer volumes test/e2e/storage/subpath.go:40 - STEP: Setting up data 12/14/22 16:47:27.161 + STEP: Setting up data 02/25/23 11:13:35.581 [It] should support subpaths with secret pod [Conformance] test/e2e/storage/subpath.go:60 - STEP: Creating pod pod-subpath-test-secret-n7f4 12/14/22 16:47:27.198 - STEP: Creating a pod to test atomic-volume-subpath 12/14/22 16:47:27.198 - Dec 14 16:47:27.217: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-n7f4" in namespace "subpath-3167" to be "Succeeded or Failed" - Dec 14 16:47:27.224: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Pending", Reason="", readiness=false. Elapsed: 6.723888ms - Dec 14 16:47:29.231: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 2.014069242s - Dec 14 16:47:31.230: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 4.013072841s - Dec 14 16:47:33.234: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 6.016520326s - Dec 14 16:47:35.231: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 8.013831662s - Dec 14 16:47:37.231: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 10.01370429s - Dec 14 16:47:39.233: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 12.015952509s - Dec 14 16:47:41.231: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 14.014035426s - Dec 14 16:47:43.230: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 16.013110031s - Dec 14 16:47:45.237: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 18.019907593s - Dec 14 16:47:47.231: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=true. Elapsed: 20.013647413s - Dec 14 16:47:49.232: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Running", Reason="", readiness=false. Elapsed: 22.01467481s - Dec 14 16:47:51.232: INFO: Pod "pod-subpath-test-secret-n7f4": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.014581866s - STEP: Saw pod success 12/14/22 16:47:51.232 - Dec 14 16:47:51.232: INFO: Pod "pod-subpath-test-secret-n7f4" satisfied condition "Succeeded or Failed" - Dec 14 16:47:51.237: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-subpath-test-secret-n7f4 container test-container-subpath-secret-n7f4: - STEP: delete the pod 12/14/22 16:47:51.248 - Dec 14 16:47:51.267: INFO: Waiting for pod pod-subpath-test-secret-n7f4 to disappear - Dec 14 16:47:51.272: INFO: Pod pod-subpath-test-secret-n7f4 no longer exists - STEP: Deleting pod pod-subpath-test-secret-n7f4 12/14/22 16:47:51.272 - Dec 14 16:47:51.272: INFO: Deleting pod "pod-subpath-test-secret-n7f4" in namespace "subpath-3167" + STEP: Creating pod pod-subpath-test-secret-rsr2 02/25/23 11:13:35.61 + STEP: Creating a pod to test atomic-volume-subpath 02/25/23 11:13:35.61 + Feb 25 11:13:35.634: INFO: Waiting up to 5m0s for pod "pod-subpath-test-secret-rsr2" in namespace "subpath-7797" to be "Succeeded or Failed" + Feb 25 11:13:35.649: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Pending", Reason="", readiness=false. Elapsed: 15.11444ms + Feb 25 11:13:37.658: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 2.024082612s + Feb 25 11:13:39.673: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 4.038289298s + Feb 25 11:13:41.658: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 6.023656371s + Feb 25 11:13:43.660: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 8.025167386s + Feb 25 11:13:45.671: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 10.036309485s + Feb 25 11:13:47.658: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 12.024031526s + Feb 25 11:13:49.665: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 14.030428174s + Feb 25 11:13:51.663: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 16.028908607s + Feb 25 11:13:53.657: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 18.022516341s + Feb 25 11:13:55.657: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=true. Elapsed: 20.022614664s + Feb 25 11:13:57.658: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Running", Reason="", readiness=false. Elapsed: 22.023371834s + Feb 25 11:13:59.659: INFO: Pod "pod-subpath-test-secret-rsr2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 24.024637256s + STEP: Saw pod success 02/25/23 11:13:59.659 + Feb 25 11:13:59.660: INFO: Pod "pod-subpath-test-secret-rsr2" satisfied condition "Succeeded or Failed" + Feb 25 11:13:59.666: INFO: Trying to get logs from node diek3dua9gay-2 pod pod-subpath-test-secret-rsr2 container test-container-subpath-secret-rsr2: + STEP: delete the pod 02/25/23 11:13:59.694 + Feb 25 11:13:59.716: INFO: Waiting for pod pod-subpath-test-secret-rsr2 to disappear + Feb 25 11:13:59.721: INFO: Pod pod-subpath-test-secret-rsr2 no longer exists + STEP: Deleting pod pod-subpath-test-secret-rsr2 02/25/23 11:13:59.721 + Feb 25 11:13:59.721: INFO: Deleting pod "pod-subpath-test-secret-rsr2" in namespace "subpath-7797" [AfterEach] [sig-storage] Subpath test/e2e/framework/framework.go:187 - Dec 14 16:47:51.276: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "subpath-3167" for this suite. 12/14/22 16:47:51.282 + Feb 25 11:13:59.726: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "subpath-7797" for this suite. 02/25/23 11:13:59.733 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSS ------------------------------ -[sig-storage] Projected secret - should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:45 -[BeforeEach] [sig-storage] Projected secret +[sig-network] Services + should find a service from listing all namespaces [Conformance] + test/e2e/network/service.go:3206 +[BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:47:51.3 -Dec 14 16:47:51.300: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 16:47:51.302 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:47:51.328 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:47:51.332 -[It] should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:45 -STEP: Creating projection with secret that has name projected-secret-test-2dd02075-6193-470e-bc91-446e5806f0e0 12/14/22 16:47:51.336 -STEP: Creating a pod to test consume secrets 12/14/22 16:47:51.345 -Dec 14 16:47:51.357: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452" in namespace "projected-453" to be "Succeeded or Failed" -Dec 14 16:47:51.364: INFO: Pod "pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452": Phase="Pending", Reason="", readiness=false. Elapsed: 6.435675ms -Dec 14 16:47:53.372: INFO: Pod "pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014631567s -Dec 14 16:47:55.372: INFO: Pod "pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014801446s -STEP: Saw pod success 12/14/22 16:47:55.372 -Dec 14 16:47:55.373: INFO: Pod "pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452" satisfied condition "Succeeded or Failed" -Dec 14 16:47:55.378: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452 container projected-secret-volume-test: -STEP: delete the pod 12/14/22 16:47:55.424 -Dec 14 16:47:55.443: INFO: Waiting for pod pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452 to disappear -Dec 14 16:47:55.448: INFO: Pod pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452 no longer exists -[AfterEach] [sig-storage] Projected secret +STEP: Creating a kubernetes client 02/25/23 11:13:59.753 +Feb 25 11:13:59.753: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 11:13:59.756 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:59.784 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:59.789 +[BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 +[It] should find a service from listing all namespaces [Conformance] + test/e2e/network/service.go:3206 +STEP: fetching services 02/25/23 11:13:59.795 +[AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 16:47:55.448: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-453" for this suite. 12/14/22 16:47:55.456 -{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance]","completed":252,"skipped":4637,"failed":0} +Feb 25 11:13:59.804: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-9850" for this suite. 02/25/23 11:13:59.811 +[AfterEach] [sig-network] Services + test/e2e/network/service.go:762 +{"msg":"PASSED [sig-network] Services should find a service from listing all namespaces [Conformance]","completed":248,"skipped":4664,"failed":0} ------------------------------ -• [4.166 seconds] -[sig-storage] Projected secret -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:45 +• [0.075 seconds] +[sig-network] Services +test/e2e/network/common/framework.go:23 + should find a service from listing all namespaces [Conformance] + test/e2e/network/service.go:3206 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected secret + [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:47:51.3 - Dec 14 16:47:51.300: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 16:47:51.302 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:47:51.328 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:47:51.332 - [It] should be consumable from pods in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:45 - STEP: Creating projection with secret that has name projected-secret-test-2dd02075-6193-470e-bc91-446e5806f0e0 12/14/22 16:47:51.336 - STEP: Creating a pod to test consume secrets 12/14/22 16:47:51.345 - Dec 14 16:47:51.357: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452" in namespace "projected-453" to be "Succeeded or Failed" - Dec 14 16:47:51.364: INFO: Pod "pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452": Phase="Pending", Reason="", readiness=false. Elapsed: 6.435675ms - Dec 14 16:47:53.372: INFO: Pod "pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014631567s - Dec 14 16:47:55.372: INFO: Pod "pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014801446s - STEP: Saw pod success 12/14/22 16:47:55.372 - Dec 14 16:47:55.373: INFO: Pod "pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452" satisfied condition "Succeeded or Failed" - Dec 14 16:47:55.378: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452 container projected-secret-volume-test: - STEP: delete the pod 12/14/22 16:47:55.424 - Dec 14 16:47:55.443: INFO: Waiting for pod pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452 to disappear - Dec 14 16:47:55.448: INFO: Pod pod-projected-secrets-6e7ea5ea-37c0-49fe-af5e-349ba4a9f452 no longer exists - [AfterEach] [sig-storage] Projected secret + STEP: Creating a kubernetes client 02/25/23 11:13:59.753 + Feb 25 11:13:59.753: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 11:13:59.756 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:59.784 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:59.789 + [BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 + [It] should find a service from listing all namespaces [Conformance] + test/e2e/network/service.go:3206 + STEP: fetching services 02/25/23 11:13:59.795 + [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 16:47:55.448: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-453" for this suite. 12/14/22 16:47:55.456 + Feb 25 11:13:59.804: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-9850" for this suite. 02/25/23 11:13:59.811 + [AfterEach] [sig-network] Services + test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSS +SSS ------------------------------ -[sig-auth] ServiceAccounts - ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance] - test/e2e/auth/service_accounts.go:528 -[BeforeEach] [sig-auth] ServiceAccounts +[sig-node] Pods + should support remote command execution over websockets [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:535 +[BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:47:55.472 -Dec 14 16:47:55.472: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename svcaccounts 12/14/22 16:47:55.474 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:47:55.516 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:47:55.521 -[It] ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance] - test/e2e/auth/service_accounts.go:528 -Dec 14 16:47:55.566: INFO: created pod -Dec 14 16:47:55.566: INFO: Waiting up to 5m0s for pod "oidc-discovery-validator" in namespace "svcaccounts-4278" to be "Succeeded or Failed" -Dec 14 16:47:55.573: INFO: Pod "oidc-discovery-validator": Phase="Pending", Reason="", readiness=false. Elapsed: 6.791572ms -Dec 14 16:47:57.581: INFO: Pod "oidc-discovery-validator": Phase="Running", Reason="", readiness=true. Elapsed: 2.015103215s -Dec 14 16:47:59.581: INFO: Pod "oidc-discovery-validator": Phase="Running", Reason="", readiness=false. Elapsed: 4.014963494s -Dec 14 16:48:01.580: INFO: Pod "oidc-discovery-validator": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.013869837s -STEP: Saw pod success 12/14/22 16:48:01.58 -Dec 14 16:48:01.581: INFO: Pod "oidc-discovery-validator" satisfied condition "Succeeded or Failed" -Dec 14 16:48:31.581: INFO: polling logs -Dec 14 16:48:31.598: INFO: Pod logs: -I1214 16:47:56.638184 1 log.go:195] OK: Got token -I1214 16:47:56.638285 1 log.go:195] validating with in-cluster discovery -I1214 16:47:56.639097 1 log.go:195] OK: got issuer https://kubernetes.default.svc.cluster.local -I1214 16:47:56.639149 1 log.go:195] Full, not-validated claims: -openidmetadata.claims{Claims:jwt.Claims{Issuer:"https://kubernetes.default.svc.cluster.local", Subject:"system:serviceaccount:svcaccounts-4278:default", Audience:jwt.Audience{"oidc-discovery-test"}, Expiry:1671037075, NotBefore:1671036475, IssuedAt:1671036475, ID:""}, Kubernetes:openidmetadata.kubeClaims{Namespace:"svcaccounts-4278", ServiceAccount:openidmetadata.kubeName{Name:"default", UID:"3e3a4182-4b4e-4720-add6-81b472919b1a"}}} -I1214 16:47:56.662241 1 log.go:195] OK: Constructed OIDC provider for issuer https://kubernetes.default.svc.cluster.local -I1214 16:47:56.670246 1 log.go:195] OK: Validated signature on JWT -I1214 16:47:56.670380 1 log.go:195] OK: Got valid claims from token! -I1214 16:47:56.670418 1 log.go:195] Full, validated claims: -&openidmetadata.claims{Claims:jwt.Claims{Issuer:"https://kubernetes.default.svc.cluster.local", Subject:"system:serviceaccount:svcaccounts-4278:default", Audience:jwt.Audience{"oidc-discovery-test"}, Expiry:1671037075, NotBefore:1671036475, IssuedAt:1671036475, ID:""}, Kubernetes:openidmetadata.kubeClaims{Namespace:"svcaccounts-4278", ServiceAccount:openidmetadata.kubeName{Name:"default", UID:"3e3a4182-4b4e-4720-add6-81b472919b1a"}}} - -Dec 14 16:48:31.598: INFO: completed pod -[AfterEach] [sig-auth] ServiceAccounts +STEP: Creating a kubernetes client 02/25/23 11:13:59.828 +Feb 25 11:13:59.829: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename pods 02/25/23 11:13:59.831 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:59.861 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:59.872 +[BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 +[It] should support remote command execution over websockets [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:535 +Feb 25 11:13:59.875: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: creating the pod 02/25/23 11:13:59.876 +STEP: submitting the pod to kubernetes 02/25/23 11:13:59.876 +Feb 25 11:13:59.893: INFO: Waiting up to 5m0s for pod "pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d" in namespace "pods-4045" to be "running and ready" +Feb 25 11:13:59.902: INFO: Pod "pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d": Phase="Pending", Reason="", readiness=false. Elapsed: 9.002312ms +Feb 25 11:13:59.902: INFO: The phase of Pod pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:14:01.913: INFO: Pod "pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019459885s +Feb 25 11:14:01.913: INFO: The phase of Pod pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:14:03.912: INFO: Pod "pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d": Phase="Running", Reason="", readiness=true. Elapsed: 4.018944659s +Feb 25 11:14:03.912: INFO: The phase of Pod pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d is Running (Ready = true) +Feb 25 11:14:03.912: INFO: Pod "pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d" satisfied condition "running and ready" +[AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 -Dec 14 16:48:31.611: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "svcaccounts-4278" for this suite. 12/14/22 16:48:31.619 -{"msg":"PASSED [sig-auth] ServiceAccounts ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance]","completed":253,"skipped":4652,"failed":0} +Feb 25 11:14:04.110: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-4045" for this suite. 02/25/23 11:14:04.12 +{"msg":"PASSED [sig-node] Pods should support remote command execution over websockets [NodeConformance] [Conformance]","completed":249,"skipped":4667,"failed":0} ------------------------------ -• [SLOW TEST] [36.156 seconds] -[sig-auth] ServiceAccounts -test/e2e/auth/framework.go:23 - ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance] - test/e2e/auth/service_accounts.go:528 +• [4.307 seconds] +[sig-node] Pods +test/e2e/common/node/framework.go:23 + should support remote command execution over websockets [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:535 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-auth] ServiceAccounts + [BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:47:55.472 - Dec 14 16:47:55.472: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename svcaccounts 12/14/22 16:47:55.474 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:47:55.516 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:47:55.521 - [It] ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance] - test/e2e/auth/service_accounts.go:528 - Dec 14 16:47:55.566: INFO: created pod - Dec 14 16:47:55.566: INFO: Waiting up to 5m0s for pod "oidc-discovery-validator" in namespace "svcaccounts-4278" to be "Succeeded or Failed" - Dec 14 16:47:55.573: INFO: Pod "oidc-discovery-validator": Phase="Pending", Reason="", readiness=false. Elapsed: 6.791572ms - Dec 14 16:47:57.581: INFO: Pod "oidc-discovery-validator": Phase="Running", Reason="", readiness=true. Elapsed: 2.015103215s - Dec 14 16:47:59.581: INFO: Pod "oidc-discovery-validator": Phase="Running", Reason="", readiness=false. Elapsed: 4.014963494s - Dec 14 16:48:01.580: INFO: Pod "oidc-discovery-validator": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.013869837s - STEP: Saw pod success 12/14/22 16:48:01.58 - Dec 14 16:48:01.581: INFO: Pod "oidc-discovery-validator" satisfied condition "Succeeded or Failed" - Dec 14 16:48:31.581: INFO: polling logs - Dec 14 16:48:31.598: INFO: Pod logs: - I1214 16:47:56.638184 1 log.go:195] OK: Got token - I1214 16:47:56.638285 1 log.go:195] validating with in-cluster discovery - I1214 16:47:56.639097 1 log.go:195] OK: got issuer https://kubernetes.default.svc.cluster.local - I1214 16:47:56.639149 1 log.go:195] Full, not-validated claims: - openidmetadata.claims{Claims:jwt.Claims{Issuer:"https://kubernetes.default.svc.cluster.local", Subject:"system:serviceaccount:svcaccounts-4278:default", Audience:jwt.Audience{"oidc-discovery-test"}, Expiry:1671037075, NotBefore:1671036475, IssuedAt:1671036475, ID:""}, Kubernetes:openidmetadata.kubeClaims{Namespace:"svcaccounts-4278", ServiceAccount:openidmetadata.kubeName{Name:"default", UID:"3e3a4182-4b4e-4720-add6-81b472919b1a"}}} - I1214 16:47:56.662241 1 log.go:195] OK: Constructed OIDC provider for issuer https://kubernetes.default.svc.cluster.local - I1214 16:47:56.670246 1 log.go:195] OK: Validated signature on JWT - I1214 16:47:56.670380 1 log.go:195] OK: Got valid claims from token! - I1214 16:47:56.670418 1 log.go:195] Full, validated claims: - &openidmetadata.claims{Claims:jwt.Claims{Issuer:"https://kubernetes.default.svc.cluster.local", Subject:"system:serviceaccount:svcaccounts-4278:default", Audience:jwt.Audience{"oidc-discovery-test"}, Expiry:1671037075, NotBefore:1671036475, IssuedAt:1671036475, ID:""}, Kubernetes:openidmetadata.kubeClaims{Namespace:"svcaccounts-4278", ServiceAccount:openidmetadata.kubeName{Name:"default", UID:"3e3a4182-4b4e-4720-add6-81b472919b1a"}}} - - Dec 14 16:48:31.598: INFO: completed pod - [AfterEach] [sig-auth] ServiceAccounts + STEP: Creating a kubernetes client 02/25/23 11:13:59.828 + Feb 25 11:13:59.829: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename pods 02/25/23 11:13:59.831 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:13:59.861 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:13:59.872 + [BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 + [It] should support remote command execution over websockets [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:535 + Feb 25 11:13:59.875: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: creating the pod 02/25/23 11:13:59.876 + STEP: submitting the pod to kubernetes 02/25/23 11:13:59.876 + Feb 25 11:13:59.893: INFO: Waiting up to 5m0s for pod "pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d" in namespace "pods-4045" to be "running and ready" + Feb 25 11:13:59.902: INFO: Pod "pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d": Phase="Pending", Reason="", readiness=false. Elapsed: 9.002312ms + Feb 25 11:13:59.902: INFO: The phase of Pod pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:14:01.913: INFO: Pod "pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019459885s + Feb 25 11:14:01.913: INFO: The phase of Pod pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:14:03.912: INFO: Pod "pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d": Phase="Running", Reason="", readiness=true. Elapsed: 4.018944659s + Feb 25 11:14:03.912: INFO: The phase of Pod pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d is Running (Ready = true) + Feb 25 11:14:03.912: INFO: Pod "pod-exec-websocket-6174d924-9f91-44b3-ab85-5dc99217218d" satisfied condition "running and ready" + [AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 - Dec 14 16:48:31.611: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "svcaccounts-4278" for this suite. 12/14/22 16:48:31.619 + Feb 25 11:14:04.110: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "pods-4045" for this suite. 02/25/23 11:14:04.12 << End Captured GinkgoWriter Output ------------------------------ SSS ------------------------------ -[sig-node] Downward API - should provide pod UID as env vars [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:266 -[BeforeEach] [sig-node] Downward API +[sig-storage] EmptyDir volumes + volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:156 +[BeforeEach] [sig-storage] EmptyDir volumes + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 11:14:04.136 +Feb 25 11:14:04.136: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir 02/25/23 11:14:04.14 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:14:04.166 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:14:04.173 +[It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:156 +STEP: Creating a pod to test emptydir volume type on node default medium 02/25/23 11:14:04.179 +Feb 25 11:14:04.197: INFO: Waiting up to 5m0s for pod "pod-e7961b9d-93fa-438f-b2fa-5139eb569438" in namespace "emptydir-8168" to be "Succeeded or Failed" +Feb 25 11:14:04.207: INFO: Pod "pod-e7961b9d-93fa-438f-b2fa-5139eb569438": Phase="Pending", Reason="", readiness=false. Elapsed: 9.807483ms +Feb 25 11:14:06.215: INFO: Pod "pod-e7961b9d-93fa-438f-b2fa-5139eb569438": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01794579s +Feb 25 11:14:08.213: INFO: Pod "pod-e7961b9d-93fa-438f-b2fa-5139eb569438": Phase="Pending", Reason="", readiness=false. Elapsed: 4.016331163s +Feb 25 11:14:10.216: INFO: Pod "pod-e7961b9d-93fa-438f-b2fa-5139eb569438": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.019175717s +STEP: Saw pod success 02/25/23 11:14:10.216 +Feb 25 11:14:10.217: INFO: Pod "pod-e7961b9d-93fa-438f-b2fa-5139eb569438" satisfied condition "Succeeded or Failed" +Feb 25 11:14:10.223: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-e7961b9d-93fa-438f-b2fa-5139eb569438 container test-container: +STEP: delete the pod 02/25/23 11:14:10.235 +Feb 25 11:14:10.265: INFO: Waiting for pod pod-e7961b9d-93fa-438f-b2fa-5139eb569438 to disappear +Feb 25 11:14:10.279: INFO: Pod pod-e7961b9d-93fa-438f-b2fa-5139eb569438 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes + test/e2e/framework/framework.go:187 +Feb 25 11:14:10.280: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-8168" for this suite. 02/25/23 11:14:10.287 +{"msg":"PASSED [sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","completed":250,"skipped":4670,"failed":0} +------------------------------ +• [SLOW TEST] [6.164 seconds] +[sig-storage] EmptyDir volumes +test/e2e/common/storage/framework.go:23 + volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:156 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-storage] EmptyDir volumes + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 11:14:04.136 + Feb 25 11:14:04.136: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir 02/25/23 11:14:04.14 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:14:04.166 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:14:04.173 + [It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:156 + STEP: Creating a pod to test emptydir volume type on node default medium 02/25/23 11:14:04.179 + Feb 25 11:14:04.197: INFO: Waiting up to 5m0s for pod "pod-e7961b9d-93fa-438f-b2fa-5139eb569438" in namespace "emptydir-8168" to be "Succeeded or Failed" + Feb 25 11:14:04.207: INFO: Pod "pod-e7961b9d-93fa-438f-b2fa-5139eb569438": Phase="Pending", Reason="", readiness=false. Elapsed: 9.807483ms + Feb 25 11:14:06.215: INFO: Pod "pod-e7961b9d-93fa-438f-b2fa-5139eb569438": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01794579s + Feb 25 11:14:08.213: INFO: Pod "pod-e7961b9d-93fa-438f-b2fa-5139eb569438": Phase="Pending", Reason="", readiness=false. Elapsed: 4.016331163s + Feb 25 11:14:10.216: INFO: Pod "pod-e7961b9d-93fa-438f-b2fa-5139eb569438": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.019175717s + STEP: Saw pod success 02/25/23 11:14:10.216 + Feb 25 11:14:10.217: INFO: Pod "pod-e7961b9d-93fa-438f-b2fa-5139eb569438" satisfied condition "Succeeded or Failed" + Feb 25 11:14:10.223: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-e7961b9d-93fa-438f-b2fa-5139eb569438 container test-container: + STEP: delete the pod 02/25/23 11:14:10.235 + Feb 25 11:14:10.265: INFO: Waiting for pod pod-e7961b9d-93fa-438f-b2fa-5139eb569438 to disappear + Feb 25 11:14:10.279: INFO: Pod pod-e7961b9d-93fa-438f-b2fa-5139eb569438 no longer exists + [AfterEach] [sig-storage] EmptyDir volumes + test/e2e/framework/framework.go:187 + Feb 25 11:14:10.280: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-8168" for this suite. 02/25/23 11:14:10.287 + << End Captured GinkgoWriter Output +------------------------------ +SSSS +------------------------------ +[sig-api-machinery] ResourceQuota + should create a ResourceQuota and capture the life of a configMap. [Conformance] + test/e2e/apimachinery/resource_quota.go:316 +[BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:48:31.629 -Dec 14 16:48:31.629: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 16:48:31.631 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:48:31.654 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:48:31.658 -[It] should provide pod UID as env vars [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:266 -STEP: Creating a pod to test downward api env vars 12/14/22 16:48:31.662 -Dec 14 16:48:31.674: INFO: Waiting up to 5m0s for pod "downward-api-e2135f1f-6573-4764-b803-b2f13fa25128" in namespace "downward-api-6389" to be "Succeeded or Failed" -Dec 14 16:48:31.678: INFO: Pod "downward-api-e2135f1f-6573-4764-b803-b2f13fa25128": Phase="Pending", Reason="", readiness=false. Elapsed: 4.315625ms -Dec 14 16:48:33.685: INFO: Pod "downward-api-e2135f1f-6573-4764-b803-b2f13fa25128": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011038696s -Dec 14 16:48:35.684: INFO: Pod "downward-api-e2135f1f-6573-4764-b803-b2f13fa25128": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010602492s -STEP: Saw pod success 12/14/22 16:48:35.685 -Dec 14 16:48:35.685: INFO: Pod "downward-api-e2135f1f-6573-4764-b803-b2f13fa25128" satisfied condition "Succeeded or Failed" -Dec 14 16:48:35.691: INFO: Trying to get logs from node kiedai4iakae-3 pod downward-api-e2135f1f-6573-4764-b803-b2f13fa25128 container dapi-container: -STEP: delete the pod 12/14/22 16:48:35.704 -Dec 14 16:48:35.721: INFO: Waiting for pod downward-api-e2135f1f-6573-4764-b803-b2f13fa25128 to disappear -Dec 14 16:48:35.725: INFO: Pod downward-api-e2135f1f-6573-4764-b803-b2f13fa25128 no longer exists -[AfterEach] [sig-node] Downward API +STEP: Creating a kubernetes client 02/25/23 11:14:10.301 +Feb 25 11:14:10.301: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename resourcequota 02/25/23 11:14:10.306 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:14:10.338 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:14:10.343 +[It] should create a ResourceQuota and capture the life of a configMap. [Conformance] + test/e2e/apimachinery/resource_quota.go:316 +STEP: Counting existing ResourceQuota 02/25/23 11:14:27.357 +STEP: Creating a ResourceQuota 02/25/23 11:14:32.366 +STEP: Ensuring resource quota status is calculated 02/25/23 11:14:32.377 +STEP: Creating a ConfigMap 02/25/23 11:14:34.386 +STEP: Ensuring resource quota status captures configMap creation 02/25/23 11:14:34.413 +STEP: Deleting a ConfigMap 02/25/23 11:14:36.422 +STEP: Ensuring resource quota status released usage 02/25/23 11:14:36.436 +[AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 -Dec 14 16:48:35.725: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-6389" for this suite. 12/14/22 16:48:35.731 -{"msg":"PASSED [sig-node] Downward API should provide pod UID as env vars [NodeConformance] [Conformance]","completed":254,"skipped":4655,"failed":0} +Feb 25 11:14:38.446: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-3588" for this suite. 02/25/23 11:14:38.457 +{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a configMap. [Conformance]","completed":251,"skipped":4674,"failed":0} ------------------------------ -• [4.111 seconds] -[sig-node] Downward API -test/e2e/common/node/framework.go:23 - should provide pod UID as env vars [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:266 +• [SLOW TEST] [28.169 seconds] +[sig-api-machinery] ResourceQuota +test/e2e/apimachinery/framework.go:23 + should create a ResourceQuota and capture the life of a configMap. [Conformance] + test/e2e/apimachinery/resource_quota.go:316 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Downward API + [BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:48:31.629 - Dec 14 16:48:31.629: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 16:48:31.631 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:48:31.654 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:48:31.658 - [It] should provide pod UID as env vars [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:266 - STEP: Creating a pod to test downward api env vars 12/14/22 16:48:31.662 - Dec 14 16:48:31.674: INFO: Waiting up to 5m0s for pod "downward-api-e2135f1f-6573-4764-b803-b2f13fa25128" in namespace "downward-api-6389" to be "Succeeded or Failed" - Dec 14 16:48:31.678: INFO: Pod "downward-api-e2135f1f-6573-4764-b803-b2f13fa25128": Phase="Pending", Reason="", readiness=false. Elapsed: 4.315625ms - Dec 14 16:48:33.685: INFO: Pod "downward-api-e2135f1f-6573-4764-b803-b2f13fa25128": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011038696s - Dec 14 16:48:35.684: INFO: Pod "downward-api-e2135f1f-6573-4764-b803-b2f13fa25128": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010602492s - STEP: Saw pod success 12/14/22 16:48:35.685 - Dec 14 16:48:35.685: INFO: Pod "downward-api-e2135f1f-6573-4764-b803-b2f13fa25128" satisfied condition "Succeeded or Failed" - Dec 14 16:48:35.691: INFO: Trying to get logs from node kiedai4iakae-3 pod downward-api-e2135f1f-6573-4764-b803-b2f13fa25128 container dapi-container: - STEP: delete the pod 12/14/22 16:48:35.704 - Dec 14 16:48:35.721: INFO: Waiting for pod downward-api-e2135f1f-6573-4764-b803-b2f13fa25128 to disappear - Dec 14 16:48:35.725: INFO: Pod downward-api-e2135f1f-6573-4764-b803-b2f13fa25128 no longer exists - [AfterEach] [sig-node] Downward API + STEP: Creating a kubernetes client 02/25/23 11:14:10.301 + Feb 25 11:14:10.301: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename resourcequota 02/25/23 11:14:10.306 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:14:10.338 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:14:10.343 + [It] should create a ResourceQuota and capture the life of a configMap. [Conformance] + test/e2e/apimachinery/resource_quota.go:316 + STEP: Counting existing ResourceQuota 02/25/23 11:14:27.357 + STEP: Creating a ResourceQuota 02/25/23 11:14:32.366 + STEP: Ensuring resource quota status is calculated 02/25/23 11:14:32.377 + STEP: Creating a ConfigMap 02/25/23 11:14:34.386 + STEP: Ensuring resource quota status captures configMap creation 02/25/23 11:14:34.413 + STEP: Deleting a ConfigMap 02/25/23 11:14:36.422 + STEP: Ensuring resource quota status released usage 02/25/23 11:14:36.436 + [AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 - Dec 14 16:48:35.725: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-6389" for this suite. 12/14/22 16:48:35.731 + Feb 25 11:14:38.446: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "resourcequota-3588" for this suite. 02/25/23 11:14:38.457 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] PodTemplates - should replace a pod template [Conformance] - test/e2e/common/node/podtemplates.go:176 -[BeforeEach] [sig-node] PodTemplates +[sig-network] Services + should serve a basic endpoint from pods [Conformance] + test/e2e/network/service.go:791 +[BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:48:35.745 -Dec 14 16:48:35.745: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename podtemplate 12/14/22 16:48:35.747 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:48:35.783 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:48:35.788 -[It] should replace a pod template [Conformance] - test/e2e/common/node/podtemplates.go:176 -STEP: Create a pod template 12/14/22 16:48:35.791 -STEP: Replace a pod template 12/14/22 16:48:35.803 -Dec 14 16:48:35.816: INFO: Found updated podtemplate annotation: "true" - -[AfterEach] [sig-node] PodTemplates +STEP: Creating a kubernetes client 02/25/23 11:14:38.472 +Feb 25 11:14:38.472: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 11:14:38.476 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:14:38.508 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:14:38.515 +[BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 +[It] should serve a basic endpoint from pods [Conformance] + test/e2e/network/service.go:791 +STEP: creating service endpoint-test2 in namespace services-2145 02/25/23 11:14:38.52 +STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2145 to expose endpoints map[] 02/25/23 11:14:38.56 +Feb 25 11:14:38.583: INFO: successfully validated that service endpoint-test2 in namespace services-2145 exposes endpoints map[] +STEP: Creating pod pod1 in namespace services-2145 02/25/23 11:14:38.583 +Feb 25 11:14:38.612: INFO: Waiting up to 5m0s for pod "pod1" in namespace "services-2145" to be "running and ready" +Feb 25 11:14:38.634: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 21.692642ms +Feb 25 11:14:38.634: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:14:40.643: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030815655s +Feb 25 11:14:40.643: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:14:42.642: INFO: Pod "pod1": Phase="Running", Reason="", readiness=true. Elapsed: 4.029332853s +Feb 25 11:14:42.642: INFO: The phase of Pod pod1 is Running (Ready = true) +Feb 25 11:14:42.642: INFO: Pod "pod1" satisfied condition "running and ready" +STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2145 to expose endpoints map[pod1:[80]] 02/25/23 11:14:42.65 +Feb 25 11:14:42.676: INFO: successfully validated that service endpoint-test2 in namespace services-2145 exposes endpoints map[pod1:[80]] +STEP: Checking if the Service forwards traffic to pod1 02/25/23 11:14:42.677 +Feb 25 11:14:42.677: INFO: Creating new exec pod +Feb 25 11:14:42.686: INFO: Waiting up to 5m0s for pod "execpodgnrd4" in namespace "services-2145" to be "running" +Feb 25 11:14:42.693: INFO: Pod "execpodgnrd4": Phase="Pending", Reason="", readiness=false. Elapsed: 7.311286ms +Feb 25 11:14:44.700: INFO: Pod "execpodgnrd4": Phase="Running", Reason="", readiness=true. Elapsed: 2.014491257s +Feb 25 11:14:44.700: INFO: Pod "execpodgnrd4" satisfied condition "running" +Feb 25 11:14:45.701: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2145 exec execpodgnrd4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 endpoint-test2 80' +Feb 25 11:14:46.327: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 endpoint-test2 80\nConnection to endpoint-test2 80 port [tcp/http] succeeded!\n" +Feb 25 11:14:46.327: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:14:46.329: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2145 exec execpodgnrd4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.24.203 80' +Feb 25 11:14:46.778: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.24.203 80\nConnection to 10.233.24.203 80 port [tcp/http] succeeded!\n" +Feb 25 11:14:46.779: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +STEP: Creating pod pod2 in namespace services-2145 02/25/23 11:14:46.779 +Feb 25 11:14:46.797: INFO: Waiting up to 5m0s for pod "pod2" in namespace "services-2145" to be "running and ready" +Feb 25 11:14:46.807: INFO: Pod "pod2": Phase="Pending", Reason="", readiness=false. Elapsed: 9.663703ms +Feb 25 11:14:46.807: INFO: The phase of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:14:48.847: INFO: Pod "pod2": Phase="Running", Reason="", readiness=true. Elapsed: 2.049185931s +Feb 25 11:14:48.847: INFO: The phase of Pod pod2 is Running (Ready = true) +Feb 25 11:14:48.847: INFO: Pod "pod2" satisfied condition "running and ready" +STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2145 to expose endpoints map[pod1:[80] pod2:[80]] 02/25/23 11:14:48.856 +Feb 25 11:14:48.937: INFO: successfully validated that service endpoint-test2 in namespace services-2145 exposes endpoints map[pod1:[80] pod2:[80]] +STEP: Checking if the Service forwards traffic to pod1 and pod2 02/25/23 11:14:48.937 +Feb 25 11:14:49.939: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2145 exec execpodgnrd4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 endpoint-test2 80' +Feb 25 11:14:50.253: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 endpoint-test2 80\nConnection to endpoint-test2 80 port [tcp/http] succeeded!\n" +Feb 25 11:14:50.253: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:14:50.254: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2145 exec execpodgnrd4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.24.203 80' +Feb 25 11:14:50.598: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.24.203 80\nConnection to 10.233.24.203 80 port [tcp/http] succeeded!\n" +Feb 25 11:14:50.598: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +STEP: Deleting pod pod1 in namespace services-2145 02/25/23 11:14:50.598 +STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2145 to expose endpoints map[pod2:[80]] 02/25/23 11:14:50.648 +Feb 25 11:14:50.682: INFO: successfully validated that service endpoint-test2 in namespace services-2145 exposes endpoints map[pod2:[80]] +STEP: Checking if the Service forwards traffic to pod2 02/25/23 11:14:50.682 +Feb 25 11:14:51.683: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2145 exec execpodgnrd4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 endpoint-test2 80' +Feb 25 11:14:52.157: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 endpoint-test2 80\nConnection to endpoint-test2 80 port [tcp/http] succeeded!\n" +Feb 25 11:14:52.157: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:14:52.157: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2145 exec execpodgnrd4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.24.203 80' +Feb 25 11:14:52.580: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.24.203 80\nConnection to 10.233.24.203 80 port [tcp/http] succeeded!\n" +Feb 25 11:14:52.580: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +STEP: Deleting pod pod2 in namespace services-2145 02/25/23 11:14:52.58 +STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2145 to expose endpoints map[] 02/25/23 11:14:52.63 +Feb 25 11:14:52.666: INFO: successfully validated that service endpoint-test2 in namespace services-2145 exposes endpoints map[] +[AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 16:48:35.816: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "podtemplate-6600" for this suite. 12/14/22 16:48:35.828 -{"msg":"PASSED [sig-node] PodTemplates should replace a pod template [Conformance]","completed":255,"skipped":4673,"failed":0} +Feb 25 11:14:52.746: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-2145" for this suite. 02/25/23 11:14:52.765 +[AfterEach] [sig-network] Services + test/e2e/network/service.go:762 +{"msg":"PASSED [sig-network] Services should serve a basic endpoint from pods [Conformance]","completed":252,"skipped":4698,"failed":0} ------------------------------ -• [0.094 seconds] -[sig-node] PodTemplates -test/e2e/common/node/framework.go:23 - should replace a pod template [Conformance] - test/e2e/common/node/podtemplates.go:176 +• [SLOW TEST] [14.361 seconds] +[sig-network] Services +test/e2e/network/common/framework.go:23 + should serve a basic endpoint from pods [Conformance] + test/e2e/network/service.go:791 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] PodTemplates + [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:48:35.745 - Dec 14 16:48:35.745: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename podtemplate 12/14/22 16:48:35.747 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:48:35.783 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:48:35.788 - [It] should replace a pod template [Conformance] - test/e2e/common/node/podtemplates.go:176 - STEP: Create a pod template 12/14/22 16:48:35.791 - STEP: Replace a pod template 12/14/22 16:48:35.803 - Dec 14 16:48:35.816: INFO: Found updated podtemplate annotation: "true" - - [AfterEach] [sig-node] PodTemplates + STEP: Creating a kubernetes client 02/25/23 11:14:38.472 + Feb 25 11:14:38.472: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 11:14:38.476 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:14:38.508 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:14:38.515 + [BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 + [It] should serve a basic endpoint from pods [Conformance] + test/e2e/network/service.go:791 + STEP: creating service endpoint-test2 in namespace services-2145 02/25/23 11:14:38.52 + STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2145 to expose endpoints map[] 02/25/23 11:14:38.56 + Feb 25 11:14:38.583: INFO: successfully validated that service endpoint-test2 in namespace services-2145 exposes endpoints map[] + STEP: Creating pod pod1 in namespace services-2145 02/25/23 11:14:38.583 + Feb 25 11:14:38.612: INFO: Waiting up to 5m0s for pod "pod1" in namespace "services-2145" to be "running and ready" + Feb 25 11:14:38.634: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 21.692642ms + Feb 25 11:14:38.634: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:14:40.643: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.030815655s + Feb 25 11:14:40.643: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:14:42.642: INFO: Pod "pod1": Phase="Running", Reason="", readiness=true. Elapsed: 4.029332853s + Feb 25 11:14:42.642: INFO: The phase of Pod pod1 is Running (Ready = true) + Feb 25 11:14:42.642: INFO: Pod "pod1" satisfied condition "running and ready" + STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2145 to expose endpoints map[pod1:[80]] 02/25/23 11:14:42.65 + Feb 25 11:14:42.676: INFO: successfully validated that service endpoint-test2 in namespace services-2145 exposes endpoints map[pod1:[80]] + STEP: Checking if the Service forwards traffic to pod1 02/25/23 11:14:42.677 + Feb 25 11:14:42.677: INFO: Creating new exec pod + Feb 25 11:14:42.686: INFO: Waiting up to 5m0s for pod "execpodgnrd4" in namespace "services-2145" to be "running" + Feb 25 11:14:42.693: INFO: Pod "execpodgnrd4": Phase="Pending", Reason="", readiness=false. Elapsed: 7.311286ms + Feb 25 11:14:44.700: INFO: Pod "execpodgnrd4": Phase="Running", Reason="", readiness=true. Elapsed: 2.014491257s + Feb 25 11:14:44.700: INFO: Pod "execpodgnrd4" satisfied condition "running" + Feb 25 11:14:45.701: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2145 exec execpodgnrd4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 endpoint-test2 80' + Feb 25 11:14:46.327: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 endpoint-test2 80\nConnection to endpoint-test2 80 port [tcp/http] succeeded!\n" + Feb 25 11:14:46.327: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:14:46.329: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2145 exec execpodgnrd4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.24.203 80' + Feb 25 11:14:46.778: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.24.203 80\nConnection to 10.233.24.203 80 port [tcp/http] succeeded!\n" + Feb 25 11:14:46.779: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + STEP: Creating pod pod2 in namespace services-2145 02/25/23 11:14:46.779 + Feb 25 11:14:46.797: INFO: Waiting up to 5m0s for pod "pod2" in namespace "services-2145" to be "running and ready" + Feb 25 11:14:46.807: INFO: Pod "pod2": Phase="Pending", Reason="", readiness=false. Elapsed: 9.663703ms + Feb 25 11:14:46.807: INFO: The phase of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:14:48.847: INFO: Pod "pod2": Phase="Running", Reason="", readiness=true. Elapsed: 2.049185931s + Feb 25 11:14:48.847: INFO: The phase of Pod pod2 is Running (Ready = true) + Feb 25 11:14:48.847: INFO: Pod "pod2" satisfied condition "running and ready" + STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2145 to expose endpoints map[pod1:[80] pod2:[80]] 02/25/23 11:14:48.856 + Feb 25 11:14:48.937: INFO: successfully validated that service endpoint-test2 in namespace services-2145 exposes endpoints map[pod1:[80] pod2:[80]] + STEP: Checking if the Service forwards traffic to pod1 and pod2 02/25/23 11:14:48.937 + Feb 25 11:14:49.939: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2145 exec execpodgnrd4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 endpoint-test2 80' + Feb 25 11:14:50.253: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 endpoint-test2 80\nConnection to endpoint-test2 80 port [tcp/http] succeeded!\n" + Feb 25 11:14:50.253: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:14:50.254: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2145 exec execpodgnrd4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.24.203 80' + Feb 25 11:14:50.598: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.24.203 80\nConnection to 10.233.24.203 80 port [tcp/http] succeeded!\n" + Feb 25 11:14:50.598: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + STEP: Deleting pod pod1 in namespace services-2145 02/25/23 11:14:50.598 + STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2145 to expose endpoints map[pod2:[80]] 02/25/23 11:14:50.648 + Feb 25 11:14:50.682: INFO: successfully validated that service endpoint-test2 in namespace services-2145 exposes endpoints map[pod2:[80]] + STEP: Checking if the Service forwards traffic to pod2 02/25/23 11:14:50.682 + Feb 25 11:14:51.683: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2145 exec execpodgnrd4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 endpoint-test2 80' + Feb 25 11:14:52.157: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 endpoint-test2 80\nConnection to endpoint-test2 80 port [tcp/http] succeeded!\n" + Feb 25 11:14:52.157: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:14:52.157: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2145 exec execpodgnrd4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.24.203 80' + Feb 25 11:14:52.580: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.24.203 80\nConnection to 10.233.24.203 80 port [tcp/http] succeeded!\n" + Feb 25 11:14:52.580: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + STEP: Deleting pod pod2 in namespace services-2145 02/25/23 11:14:52.58 + STEP: waiting up to 3m0s for service endpoint-test2 in namespace services-2145 to expose endpoints map[] 02/25/23 11:14:52.63 + Feb 25 11:14:52.666: INFO: successfully validated that service endpoint-test2 in namespace services-2145 exposes endpoints map[] + [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 16:48:35.816: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "podtemplate-6600" for this suite. 12/14/22 16:48:35.828 + Feb 25 11:14:52.746: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-2145" for this suite. 02/25/23 11:14:52.765 + [AfterEach] [sig-network] Services + test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSS ------------------------------ -[sig-node] Kubelet when scheduling a read only busybox container - should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:184 -[BeforeEach] [sig-node] Kubelet +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + should be able to deny attaching pod [Conformance] + test/e2e/apimachinery/webhook.go:208 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:48:35.841 -Dec 14 16:48:35.841: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubelet-test 12/14/22 16:48:35.844 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:48:35.868 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:48:35.876 -[BeforeEach] [sig-node] Kubelet - test/e2e/common/node/kubelet.go:41 -[It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:184 -Dec 14 16:48:35.900: INFO: Waiting up to 5m0s for pod "busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c" in namespace "kubelet-test-3788" to be "running and ready" -Dec 14 16:48:35.907: INFO: Pod "busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c": Phase="Pending", Reason="", readiness=false. Elapsed: 7.101983ms -Dec 14 16:48:35.907: INFO: The phase of Pod busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:48:37.915: INFO: Pod "busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014799216s -Dec 14 16:48:37.915: INFO: The phase of Pod busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:48:39.914: INFO: Pod "busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c": Phase="Running", Reason="", readiness=true. Elapsed: 4.014598967s -Dec 14 16:48:39.914: INFO: The phase of Pod busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c is Running (Ready = true) -Dec 14 16:48:39.914: INFO: Pod "busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c" satisfied condition "running and ready" -[AfterEach] [sig-node] Kubelet +STEP: Creating a kubernetes client 02/25/23 11:14:52.852 +Feb 25 11:14:52.856: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 11:14:52.867 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:14:52.963 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:14:52.978 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 +STEP: Setting up server cert 02/25/23 11:14:53.278 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 11:14:53.804 +STEP: Deploying the webhook pod 02/25/23 11:14:53.817 +STEP: Wait for the deployment to be ready 02/25/23 11:14:53.862 +Feb 25 11:14:53.892: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set +Feb 25 11:14:55.911: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 11, 14, 53, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 14, 53, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 11, 14, 54, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 14, 53, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} +STEP: Deploying the webhook service 02/25/23 11:14:57.921 +STEP: Verifying the service has paired with the endpoint 02/25/23 11:14:57.96 +Feb 25 11:14:58.961: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should be able to deny attaching pod [Conformance] + test/e2e/apimachinery/webhook.go:208 +STEP: Registering the webhook via the AdmissionRegistration API 02/25/23 11:14:58.967 +Feb 25 11:14:58.998: INFO: Waiting for webhook configuration to be ready... +STEP: create a pod 02/25/23 11:14:59.117 +Feb 25 11:14:59.130: INFO: Waiting up to 5m0s for pod "to-be-attached-pod" in namespace "webhook-9519" to be "running" +Feb 25 11:14:59.141: INFO: Pod "to-be-attached-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 10.725781ms +Feb 25 11:15:01.147: INFO: Pod "to-be-attached-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.017024836s +Feb 25 11:15:01.147: INFO: Pod "to-be-attached-pod" satisfied condition "running" +STEP: 'kubectl attach' the pod, should be denied by the webhook 02/25/23 11:15:01.148 +Feb 25 11:15:01.148: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=webhook-9519 attach --namespace=webhook-9519 to-be-attached-pod -i -c=container1' +Feb 25 11:15:01.398: INFO: rc: 1 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:48:40.107: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubelet-test-3788" for this suite. 12/14/22 16:48:40.116 -{"msg":"PASSED [sig-node] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]","completed":256,"skipped":4676,"failed":0} +Feb 25 11:15:01.417: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-9519" for this suite. 02/25/23 11:15:01.431 +STEP: Destroying namespace "webhook-9519-markers" for this suite. 02/25/23 11:15:01.446 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny attaching pod [Conformance]","completed":253,"skipped":4704,"failed":0} ------------------------------ -• [4.285 seconds] -[sig-node] Kubelet -test/e2e/common/node/framework.go:23 - when scheduling a read only busybox container - test/e2e/common/node/kubelet.go:175 - should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:184 +• [SLOW TEST] [8.794 seconds] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + should be able to deny attaching pod [Conformance] + test/e2e/apimachinery/webhook.go:208 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Kubelet + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:48:35.841 - Dec 14 16:48:35.841: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubelet-test 12/14/22 16:48:35.844 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:48:35.868 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:48:35.876 - [BeforeEach] [sig-node] Kubelet - test/e2e/common/node/kubelet.go:41 - [It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/kubelet.go:184 - Dec 14 16:48:35.900: INFO: Waiting up to 5m0s for pod "busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c" in namespace "kubelet-test-3788" to be "running and ready" - Dec 14 16:48:35.907: INFO: Pod "busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c": Phase="Pending", Reason="", readiness=false. Elapsed: 7.101983ms - Dec 14 16:48:35.907: INFO: The phase of Pod busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:48:37.915: INFO: Pod "busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014799216s - Dec 14 16:48:37.915: INFO: The phase of Pod busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:48:39.914: INFO: Pod "busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c": Phase="Running", Reason="", readiness=true. Elapsed: 4.014598967s - Dec 14 16:48:39.914: INFO: The phase of Pod busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c is Running (Ready = true) - Dec 14 16:48:39.914: INFO: Pod "busybox-readonly-fs83ae722f-3299-4547-9397-d57d3ce5e45c" satisfied condition "running and ready" - [AfterEach] [sig-node] Kubelet + STEP: Creating a kubernetes client 02/25/23 11:14:52.852 + Feb 25 11:14:52.856: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 11:14:52.867 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:14:52.963 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:14:52.978 + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 + STEP: Setting up server cert 02/25/23 11:14:53.278 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 11:14:53.804 + STEP: Deploying the webhook pod 02/25/23 11:14:53.817 + STEP: Wait for the deployment to be ready 02/25/23 11:14:53.862 + Feb 25 11:14:53.892: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set + Feb 25 11:14:55.911: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 11, 14, 53, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 14, 53, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 11, 14, 54, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 14, 53, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} + STEP: Deploying the webhook service 02/25/23 11:14:57.921 + STEP: Verifying the service has paired with the endpoint 02/25/23 11:14:57.96 + Feb 25 11:14:58.961: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] should be able to deny attaching pod [Conformance] + test/e2e/apimachinery/webhook.go:208 + STEP: Registering the webhook via the AdmissionRegistration API 02/25/23 11:14:58.967 + Feb 25 11:14:58.998: INFO: Waiting for webhook configuration to be ready... + STEP: create a pod 02/25/23 11:14:59.117 + Feb 25 11:14:59.130: INFO: Waiting up to 5m0s for pod "to-be-attached-pod" in namespace "webhook-9519" to be "running" + Feb 25 11:14:59.141: INFO: Pod "to-be-attached-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 10.725781ms + Feb 25 11:15:01.147: INFO: Pod "to-be-attached-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.017024836s + Feb 25 11:15:01.147: INFO: Pod "to-be-attached-pod" satisfied condition "running" + STEP: 'kubectl attach' the pod, should be denied by the webhook 02/25/23 11:15:01.148 + Feb 25 11:15:01.148: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=webhook-9519 attach --namespace=webhook-9519 to-be-attached-pod -i -c=container1' + Feb 25 11:15:01.398: INFO: rc: 1 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:48:40.107: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubelet-test-3788" for this suite. 12/14/22 16:48:40.116 + Feb 25 11:15:01.417: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-9519" for this suite. 02/25/23 11:15:01.431 + STEP: Destroying namespace "webhook-9519-markers" for this suite. 02/25/23 11:15:01.446 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSS +SSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] ResourceQuota - should manage the lifecycle of a ResourceQuota [Conformance] - test/e2e/apimachinery/resource_quota.go:933 + should create a ResourceQuota and capture the life of a service. [Conformance] + test/e2e/apimachinery/resource_quota.go:90 [BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:48:40.128 -Dec 14 16:48:40.128: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename resourcequota 12/14/22 16:48:40.131 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:48:40.152 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:48:40.156 -[It] should manage the lifecycle of a ResourceQuota [Conformance] - test/e2e/apimachinery/resource_quota.go:933 -STEP: Creating a ResourceQuota 12/14/22 16:48:40.159 -STEP: Getting a ResourceQuota 12/14/22 16:48:40.169 -STEP: Listing all ResourceQuotas with LabelSelector 12/14/22 16:48:40.176 -STEP: Patching the ResourceQuota 12/14/22 16:48:40.18 -STEP: Deleting a Collection of ResourceQuotas 12/14/22 16:48:40.193 -STEP: Verifying the deleted ResourceQuota 12/14/22 16:48:40.207 +STEP: Creating a kubernetes client 02/25/23 11:15:01.664 +Feb 25 11:15:01.664: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename resourcequota 02/25/23 11:15:01.668 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:15:01.722 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:15:01.736 +[It] should create a ResourceQuota and capture the life of a service. [Conformance] + test/e2e/apimachinery/resource_quota.go:90 +STEP: Counting existing ResourceQuota 02/25/23 11:15:01.75 +STEP: Creating a ResourceQuota 02/25/23 11:15:06.76 +STEP: Ensuring resource quota status is calculated 02/25/23 11:15:06.783 +STEP: Creating a Service 02/25/23 11:15:08.792 +STEP: Creating a NodePort Service 02/25/23 11:15:08.829 +STEP: Not allowing a LoadBalancer Service with NodePort to be created that exceeds remaining quota 02/25/23 11:15:09.039 +STEP: Ensuring resource quota status captures service creation 02/25/23 11:15:09.351 +STEP: Deleting Services 02/25/23 11:15:11.361 +STEP: Ensuring resource quota status released usage 02/25/23 11:15:11.507 [AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 -Dec 14 16:48:40.212: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-1985" for this suite. 12/14/22 16:48:40.22 -{"msg":"PASSED [sig-api-machinery] ResourceQuota should manage the lifecycle of a ResourceQuota [Conformance]","completed":257,"skipped":4684,"failed":0} +Feb 25 11:15:13.518: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-6773" for this suite. 02/25/23 11:15:13.538 +{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a service. [Conformance]","completed":254,"skipped":4721,"failed":0} ------------------------------ -• [0.103 seconds] +• [SLOW TEST] [11.887 seconds] [sig-api-machinery] ResourceQuota test/e2e/apimachinery/framework.go:23 - should manage the lifecycle of a ResourceQuota [Conformance] - test/e2e/apimachinery/resource_quota.go:933 + should create a ResourceQuota and capture the life of a service. [Conformance] + test/e2e/apimachinery/resource_quota.go:90 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:48:40.128 - Dec 14 16:48:40.128: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename resourcequota 12/14/22 16:48:40.131 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:48:40.152 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:48:40.156 - [It] should manage the lifecycle of a ResourceQuota [Conformance] - test/e2e/apimachinery/resource_quota.go:933 - STEP: Creating a ResourceQuota 12/14/22 16:48:40.159 - STEP: Getting a ResourceQuota 12/14/22 16:48:40.169 - STEP: Listing all ResourceQuotas with LabelSelector 12/14/22 16:48:40.176 - STEP: Patching the ResourceQuota 12/14/22 16:48:40.18 - STEP: Deleting a Collection of ResourceQuotas 12/14/22 16:48:40.193 - STEP: Verifying the deleted ResourceQuota 12/14/22 16:48:40.207 + STEP: Creating a kubernetes client 02/25/23 11:15:01.664 + Feb 25 11:15:01.664: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename resourcequota 02/25/23 11:15:01.668 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:15:01.722 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:15:01.736 + [It] should create a ResourceQuota and capture the life of a service. [Conformance] + test/e2e/apimachinery/resource_quota.go:90 + STEP: Counting existing ResourceQuota 02/25/23 11:15:01.75 + STEP: Creating a ResourceQuota 02/25/23 11:15:06.76 + STEP: Ensuring resource quota status is calculated 02/25/23 11:15:06.783 + STEP: Creating a Service 02/25/23 11:15:08.792 + STEP: Creating a NodePort Service 02/25/23 11:15:08.829 + STEP: Not allowing a LoadBalancer Service with NodePort to be created that exceeds remaining quota 02/25/23 11:15:09.039 + STEP: Ensuring resource quota status captures service creation 02/25/23 11:15:09.351 + STEP: Deleting Services 02/25/23 11:15:11.361 + STEP: Ensuring resource quota status released usage 02/25/23 11:15:11.507 [AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 - Dec 14 16:48:40.212: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "resourcequota-1985" for this suite. 12/14/22 16:48:40.22 + Feb 25 11:15:13.518: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "resourcequota-6773" for this suite. 02/25/23 11:15:13.538 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Projected configMap - optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:173 -[BeforeEach] [sig-storage] Projected configMap +[sig-apps] CronJob + should replace jobs when ReplaceConcurrent [Conformance] + test/e2e/apps/cronjob.go:160 +[BeforeEach] [sig-apps] CronJob test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:48:40.237 -Dec 14 16:48:40.237: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 16:48:40.239 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:48:40.259 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:48:40.262 -[It] optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:173 -STEP: Creating configMap with name cm-test-opt-del-6ae5a9c4-beee-4653-854e-55185291a6d5 12/14/22 16:48:40.274 -STEP: Creating configMap with name cm-test-opt-upd-2b0a93bf-416f-4fac-a1e8-d2f2e298a5eb 12/14/22 16:48:40.283 -STEP: Creating the pod 12/14/22 16:48:40.29 -Dec 14 16:48:40.302: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75" in namespace "projected-1479" to be "running and ready" -Dec 14 16:48:40.311: INFO: Pod "pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75": Phase="Pending", Reason="", readiness=false. Elapsed: 9.565274ms -Dec 14 16:48:40.311: INFO: The phase of Pod pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:48:42.318: INFO: Pod "pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016019538s -Dec 14 16:48:42.318: INFO: The phase of Pod pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:48:44.319: INFO: Pod "pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75": Phase="Running", Reason="", readiness=true. Elapsed: 4.0168071s -Dec 14 16:48:44.319: INFO: The phase of Pod pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75 is Running (Ready = true) -Dec 14 16:48:44.319: INFO: Pod "pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75" satisfied condition "running and ready" -STEP: Deleting configmap cm-test-opt-del-6ae5a9c4-beee-4653-854e-55185291a6d5 12/14/22 16:48:44.377 -STEP: Updating configmap cm-test-opt-upd-2b0a93bf-416f-4fac-a1e8-d2f2e298a5eb 12/14/22 16:48:44.388 -STEP: Creating configMap with name cm-test-opt-create-f9ef762a-fa00-4bf1-9ae2-bb1f4ce20fb8 12/14/22 16:48:44.394 -STEP: waiting to observe update in volume 12/14/22 16:48:44.401 -[AfterEach] [sig-storage] Projected configMap +STEP: Creating a kubernetes client 02/25/23 11:15:13.56 +Feb 25 11:15:13.560: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename cronjob 02/25/23 11:15:13.563 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:15:13.591 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:15:13.6 +[It] should replace jobs when ReplaceConcurrent [Conformance] + test/e2e/apps/cronjob.go:160 +STEP: Creating a ReplaceConcurrent cronjob 02/25/23 11:15:13.606 +STEP: Ensuring a job is scheduled 02/25/23 11:15:13.624 +STEP: Ensuring exactly one is scheduled 02/25/23 11:16:01.631 +STEP: Ensuring exactly one running job exists by listing jobs explicitly 02/25/23 11:16:01.636 +STEP: Ensuring the job is replaced with a new one 02/25/23 11:16:01.641 +STEP: Removing cronjob 02/25/23 11:17:01.677 +[AfterEach] [sig-apps] CronJob test/e2e/framework/framework.go:187 -Dec 14 16:49:52.875: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-1479" for this suite. 12/14/22 16:49:52.884 -{"msg":"PASSED [sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance]","completed":258,"skipped":4700,"failed":0} +Feb 25 11:17:01.694: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "cronjob-8054" for this suite. 02/25/23 11:17:01.704 +{"msg":"PASSED [sig-apps] CronJob should replace jobs when ReplaceConcurrent [Conformance]","completed":255,"skipped":4764,"failed":0} ------------------------------ -• [SLOW TEST] [72.661 seconds] -[sig-storage] Projected configMap -test/e2e/common/storage/framework.go:23 - optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:173 +• [SLOW TEST] [108.161 seconds] +[sig-apps] CronJob +test/e2e/apps/framework.go:23 + should replace jobs when ReplaceConcurrent [Conformance] + test/e2e/apps/cronjob.go:160 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected configMap + [BeforeEach] [sig-apps] CronJob test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:48:40.237 - Dec 14 16:48:40.237: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 16:48:40.239 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:48:40.259 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:48:40.262 - [It] optional updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:173 - STEP: Creating configMap with name cm-test-opt-del-6ae5a9c4-beee-4653-854e-55185291a6d5 12/14/22 16:48:40.274 - STEP: Creating configMap with name cm-test-opt-upd-2b0a93bf-416f-4fac-a1e8-d2f2e298a5eb 12/14/22 16:48:40.283 - STEP: Creating the pod 12/14/22 16:48:40.29 - Dec 14 16:48:40.302: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75" in namespace "projected-1479" to be "running and ready" - Dec 14 16:48:40.311: INFO: Pod "pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75": Phase="Pending", Reason="", readiness=false. Elapsed: 9.565274ms - Dec 14 16:48:40.311: INFO: The phase of Pod pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:48:42.318: INFO: Pod "pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016019538s - Dec 14 16:48:42.318: INFO: The phase of Pod pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:48:44.319: INFO: Pod "pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75": Phase="Running", Reason="", readiness=true. Elapsed: 4.0168071s - Dec 14 16:48:44.319: INFO: The phase of Pod pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75 is Running (Ready = true) - Dec 14 16:48:44.319: INFO: Pod "pod-projected-configmaps-64b82a31-03c8-4960-a42a-6a6768073a75" satisfied condition "running and ready" - STEP: Deleting configmap cm-test-opt-del-6ae5a9c4-beee-4653-854e-55185291a6d5 12/14/22 16:48:44.377 - STEP: Updating configmap cm-test-opt-upd-2b0a93bf-416f-4fac-a1e8-d2f2e298a5eb 12/14/22 16:48:44.388 - STEP: Creating configMap with name cm-test-opt-create-f9ef762a-fa00-4bf1-9ae2-bb1f4ce20fb8 12/14/22 16:48:44.394 - STEP: waiting to observe update in volume 12/14/22 16:48:44.401 - [AfterEach] [sig-storage] Projected configMap + STEP: Creating a kubernetes client 02/25/23 11:15:13.56 + Feb 25 11:15:13.560: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename cronjob 02/25/23 11:15:13.563 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:15:13.591 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:15:13.6 + [It] should replace jobs when ReplaceConcurrent [Conformance] + test/e2e/apps/cronjob.go:160 + STEP: Creating a ReplaceConcurrent cronjob 02/25/23 11:15:13.606 + STEP: Ensuring a job is scheduled 02/25/23 11:15:13.624 + STEP: Ensuring exactly one is scheduled 02/25/23 11:16:01.631 + STEP: Ensuring exactly one running job exists by listing jobs explicitly 02/25/23 11:16:01.636 + STEP: Ensuring the job is replaced with a new one 02/25/23 11:16:01.641 + STEP: Removing cronjob 02/25/23 11:17:01.677 + [AfterEach] [sig-apps] CronJob test/e2e/framework/framework.go:187 - Dec 14 16:49:52.875: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-1479" for this suite. 12/14/22 16:49:52.884 + Feb 25 11:17:01.694: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "cronjob-8054" for this suite. 02/25/23 11:17:01.704 << End Captured GinkgoWriter Output ------------------------------ -[sig-apps] Deployment - RecreateDeployment should delete old pods and create new ones [Conformance] - test/e2e/apps/deployment.go:113 -[BeforeEach] [sig-apps] Deployment +SSSSSSS +------------------------------ +[sig-api-machinery] ResourceQuota + should be able to update and delete ResourceQuota. [Conformance] + test/e2e/apimachinery/resource_quota.go:874 +[BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:49:52.898 -Dec 14 16:49:52.898: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename deployment 12/14/22 16:49:52.901 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:49:52.932 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:49:52.936 -[BeforeEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:91 -[It] RecreateDeployment should delete old pods and create new ones [Conformance] - test/e2e/apps/deployment.go:113 -Dec 14 16:49:52.941: INFO: Creating deployment "test-recreate-deployment" -Dec 14 16:49:52.950: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1 -Dec 14 16:49:52.959: INFO: new replicaset for deployment "test-recreate-deployment" is yet to be created -Dec 14 16:49:54.972: INFO: Waiting deployment "test-recreate-deployment" to complete -Dec 14 16:49:54.977: INFO: Triggering a new rollout for deployment "test-recreate-deployment" -Dec 14 16:49:54.993: INFO: Updating deployment test-recreate-deployment -Dec 14 16:49:54.993: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods -[AfterEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:84 -Dec 14 16:49:55.272: INFO: Deployment "test-recreate-deployment": -&Deployment{ObjectMeta:{test-recreate-deployment deployment-4531 f2f22cce-c3cd-416b-a5de-eea5aafde3d8 25937 2 2022-12-14 16:49:52 +0000 UTC map[name:sample-pod-3] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2022-12-14 16:49:54 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 16:49:55 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:replicas":{},"f:unavailableReplicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003681cf8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2022-12-14 16:49:55 +0000 UTC,LastTransitionTime:2022-12-14 16:49:55 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "test-recreate-deployment-9d58999df" is progressing.,LastUpdateTime:2022-12-14 16:49:55 +0000 UTC,LastTransitionTime:2022-12-14 16:49:52 +0000 UTC,},},ReadyReplicas:0,CollisionCount:nil,},} - -Dec 14 16:49:55.277: INFO: New ReplicaSet "test-recreate-deployment-9d58999df" of Deployment "test-recreate-deployment": -&ReplicaSet{ObjectMeta:{test-recreate-deployment-9d58999df deployment-4531 a66a49b2-ea03-4aa2-8c87-a9c965d5d257 25935 1 2022-12-14 16:49:55 +0000 UTC map[name:sample-pod-3 pod-template-hash:9d58999df] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-recreate-deployment f2f22cce-c3cd-416b-a5de-eea5aafde3d8 0xc0033bc6e0 0xc0033bc6e1}] [] [{kube-controller-manager Update apps/v1 2022-12-14 16:49:55 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f2f22cce-c3cd-416b-a5de-eea5aafde3d8\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 16:49:55 +0000 UTC FieldsV1 {"f:status":{"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 9d58999df,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:9d58999df] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0033bc778 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} -Dec 14 16:49:55.277: INFO: All old ReplicaSets of Deployment "test-recreate-deployment": -Dec 14 16:49:55.278: INFO: &ReplicaSet{ObjectMeta:{test-recreate-deployment-7d8b6f647f deployment-4531 c2d4960e-0b56-43a4-b453-365843654118 25925 2 2022-12-14 16:49:52 +0000 UTC map[name:sample-pod-3 pod-template-hash:7d8b6f647f] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-recreate-deployment f2f22cce-c3cd-416b-a5de-eea5aafde3d8 0xc0033bc5c7 0xc0033bc5c8}] [] [{kube-controller-manager Update apps/v1 2022-12-14 16:49:54 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f2f22cce-c3cd-416b-a5de-eea5aafde3d8\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 16:49:55 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 7d8b6f647f,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:7d8b6f647f] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0033bc678 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} -Dec 14 16:49:55.282: INFO: Pod "test-recreate-deployment-9d58999df-ts465" is not available: -&Pod{ObjectMeta:{test-recreate-deployment-9d58999df-ts465 test-recreate-deployment-9d58999df- deployment-4531 17f31c5a-8441-4888-a6aa-5cb5f2f79b4f 25936 0 2022-12-14 16:49:55 +0000 UTC map[name:sample-pod-3 pod-template-hash:9d58999df] map[] [{apps/v1 ReplicaSet test-recreate-deployment-9d58999df a66a49b2-ea03-4aa2-8c87-a9c965d5d257 0xc0033bcbf0 0xc0033bcbf1}] [] [{kube-controller-manager Update v1 2022-12-14 16:49:55 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"a66a49b2-ea03-4aa2-8c87-a9c965d5d257\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:49:55 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-btgnn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-btgnn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:49:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:49:55 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:49:55 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:49:55 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:,StartTime:2022-12-14 16:49:55 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -[AfterEach] [sig-apps] Deployment +STEP: Creating a kubernetes client 02/25/23 11:17:01.723 +Feb 25 11:17:01.723: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename resourcequota 02/25/23 11:17:01.731 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:17:01.767 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:17:01.78 +[It] should be able to update and delete ResourceQuota. [Conformance] + test/e2e/apimachinery/resource_quota.go:874 +STEP: Creating a ResourceQuota 02/25/23 11:17:01.786 +STEP: Getting a ResourceQuota 02/25/23 11:17:01.793 +STEP: Updating a ResourceQuota 02/25/23 11:17:01.804 +STEP: Verifying a ResourceQuota was modified 02/25/23 11:17:01.816 +STEP: Deleting a ResourceQuota 02/25/23 11:17:01.834 +STEP: Verifying the deleted ResourceQuota 02/25/23 11:17:01.85 +[AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 -Dec 14 16:49:55.282: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "deployment-4531" for this suite. 12/14/22 16:49:55.287 -{"msg":"PASSED [sig-apps] Deployment RecreateDeployment should delete old pods and create new ones [Conformance]","completed":259,"skipped":4700,"failed":0} +Feb 25 11:17:01.856: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-4014" for this suite. 02/25/23 11:17:01.864 +{"msg":"PASSED [sig-api-machinery] ResourceQuota should be able to update and delete ResourceQuota. [Conformance]","completed":256,"skipped":4771,"failed":0} ------------------------------ -• [2.398 seconds] -[sig-apps] Deployment -test/e2e/apps/framework.go:23 - RecreateDeployment should delete old pods and create new ones [Conformance] - test/e2e/apps/deployment.go:113 +• [0.170 seconds] +[sig-api-machinery] ResourceQuota +test/e2e/apimachinery/framework.go:23 + should be able to update and delete ResourceQuota. [Conformance] + test/e2e/apimachinery/resource_quota.go:874 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Deployment + [BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:49:52.898 - Dec 14 16:49:52.898: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename deployment 12/14/22 16:49:52.901 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:49:52.932 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:49:52.936 - [BeforeEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:91 - [It] RecreateDeployment should delete old pods and create new ones [Conformance] - test/e2e/apps/deployment.go:113 - Dec 14 16:49:52.941: INFO: Creating deployment "test-recreate-deployment" - Dec 14 16:49:52.950: INFO: Waiting deployment "test-recreate-deployment" to be updated to revision 1 - Dec 14 16:49:52.959: INFO: new replicaset for deployment "test-recreate-deployment" is yet to be created - Dec 14 16:49:54.972: INFO: Waiting deployment "test-recreate-deployment" to complete - Dec 14 16:49:54.977: INFO: Triggering a new rollout for deployment "test-recreate-deployment" - Dec 14 16:49:54.993: INFO: Updating deployment test-recreate-deployment - Dec 14 16:49:54.993: INFO: Watching deployment "test-recreate-deployment" to verify that new pods will not run with olds pods - [AfterEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:84 - Dec 14 16:49:55.272: INFO: Deployment "test-recreate-deployment": - &Deployment{ObjectMeta:{test-recreate-deployment deployment-4531 f2f22cce-c3cd-416b-a5de-eea5aafde3d8 25937 2 2022-12-14 16:49:52 +0000 UTC map[name:sample-pod-3] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2022-12-14 16:49:54 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 16:49:55 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:replicas":{},"f:unavailableReplicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003681cf8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:Recreate,RollingUpdate:nil,},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:0,UnavailableReplicas:1,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:False,Reason:MinimumReplicasUnavailable,Message:Deployment does not have minimum availability.,LastUpdateTime:2022-12-14 16:49:55 +0000 UTC,LastTransitionTime:2022-12-14 16:49:55 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:ReplicaSetUpdated,Message:ReplicaSet "test-recreate-deployment-9d58999df" is progressing.,LastUpdateTime:2022-12-14 16:49:55 +0000 UTC,LastTransitionTime:2022-12-14 16:49:52 +0000 UTC,},},ReadyReplicas:0,CollisionCount:nil,},} - - Dec 14 16:49:55.277: INFO: New ReplicaSet "test-recreate-deployment-9d58999df" of Deployment "test-recreate-deployment": - &ReplicaSet{ObjectMeta:{test-recreate-deployment-9d58999df deployment-4531 a66a49b2-ea03-4aa2-8c87-a9c965d5d257 25935 1 2022-12-14 16:49:55 +0000 UTC map[name:sample-pod-3 pod-template-hash:9d58999df] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-recreate-deployment f2f22cce-c3cd-416b-a5de-eea5aafde3d8 0xc0033bc6e0 0xc0033bc6e1}] [] [{kube-controller-manager Update apps/v1 2022-12-14 16:49:55 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f2f22cce-c3cd-416b-a5de-eea5aafde3d8\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 16:49:55 +0000 UTC FieldsV1 {"f:status":{"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 9d58999df,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:9d58999df] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0033bc778 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:1,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} - Dec 14 16:49:55.277: INFO: All old ReplicaSets of Deployment "test-recreate-deployment": - Dec 14 16:49:55.278: INFO: &ReplicaSet{ObjectMeta:{test-recreate-deployment-7d8b6f647f deployment-4531 c2d4960e-0b56-43a4-b453-365843654118 25925 2 2022-12-14 16:49:52 +0000 UTC map[name:sample-pod-3 pod-template-hash:7d8b6f647f] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:1 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-recreate-deployment f2f22cce-c3cd-416b-a5de-eea5aafde3d8 0xc0033bc5c7 0xc0033bc5c8}] [] [{kube-controller-manager Update apps/v1 2022-12-14 16:49:54 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"f2f22cce-c3cd-416b-a5de-eea5aafde3d8\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 16:49:55 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: sample-pod-3,pod-template-hash: 7d8b6f647f,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:sample-pod-3 pod-template-hash:7d8b6f647f] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0033bc678 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} - Dec 14 16:49:55.282: INFO: Pod "test-recreate-deployment-9d58999df-ts465" is not available: - &Pod{ObjectMeta:{test-recreate-deployment-9d58999df-ts465 test-recreate-deployment-9d58999df- deployment-4531 17f31c5a-8441-4888-a6aa-5cb5f2f79b4f 25936 0 2022-12-14 16:49:55 +0000 UTC map[name:sample-pod-3 pod-template-hash:9d58999df] map[] [{apps/v1 ReplicaSet test-recreate-deployment-9d58999df a66a49b2-ea03-4aa2-8c87-a9c965d5d257 0xc0033bcbf0 0xc0033bcbf1}] [] [{kube-controller-manager Update v1 2022-12-14 16:49:55 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"a66a49b2-ea03-4aa2-8c87-a9c965d5d257\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 16:49:55 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-btgnn,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-btgnn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:49:55 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:49:55 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:49:55 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 16:49:55 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:,StartTime:2022-12-14 16:49:55 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - [AfterEach] [sig-apps] Deployment + STEP: Creating a kubernetes client 02/25/23 11:17:01.723 + Feb 25 11:17:01.723: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename resourcequota 02/25/23 11:17:01.731 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:17:01.767 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:17:01.78 + [It] should be able to update and delete ResourceQuota. [Conformance] + test/e2e/apimachinery/resource_quota.go:874 + STEP: Creating a ResourceQuota 02/25/23 11:17:01.786 + STEP: Getting a ResourceQuota 02/25/23 11:17:01.793 + STEP: Updating a ResourceQuota 02/25/23 11:17:01.804 + STEP: Verifying a ResourceQuota was modified 02/25/23 11:17:01.816 + STEP: Deleting a ResourceQuota 02/25/23 11:17:01.834 + STEP: Verifying the deleted ResourceQuota 02/25/23 11:17:01.85 + [AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 - Dec 14 16:49:55.282: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "deployment-4531" for this suite. 12/14/22 16:49:55.287 + Feb 25 11:17:01.856: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "resourcequota-4014" for this suite. 02/25/23 11:17:01.864 << End Captured GinkgoWriter Output ------------------------------ -S +SSSSSSSSSSS ------------------------------ -[sig-storage] Projected configMap - should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:88 -[BeforeEach] [sig-storage] Projected configMap +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + listing validating webhooks should work [Conformance] + test/e2e/apimachinery/webhook.go:581 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:49:55.299 -Dec 14 16:49:55.300: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 16:49:55.302 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:49:55.328 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:49:55.332 -[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:88 -STEP: Creating configMap with name projected-configmap-test-volume-map-c8be328c-ea6d-44e8-9bbe-e5a757d39ecb 12/14/22 16:49:55.334 -STEP: Creating a pod to test consume configMaps 12/14/22 16:49:55.34 -Dec 14 16:49:55.349: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d" in namespace "projected-4979" to be "Succeeded or Failed" -Dec 14 16:49:55.353: INFO: Pod "pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d": Phase="Pending", Reason="", readiness=false. Elapsed: 3.407895ms -Dec 14 16:49:57.360: INFO: Pod "pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d": Phase="Running", Reason="", readiness=true. Elapsed: 2.010458516s -Dec 14 16:49:59.359: INFO: Pod "pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d": Phase="Running", Reason="", readiness=false. Elapsed: 4.009370593s -Dec 14 16:50:01.358: INFO: Pod "pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.008575104s -STEP: Saw pod success 12/14/22 16:50:01.358 -Dec 14 16:50:01.358: INFO: Pod "pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d" satisfied condition "Succeeded or Failed" -Dec 14 16:50:01.364: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d container agnhost-container: -STEP: delete the pod 12/14/22 16:50:01.375 -Dec 14 16:50:01.390: INFO: Waiting for pod pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d to disappear -Dec 14 16:50:01.395: INFO: Pod pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d no longer exists -[AfterEach] [sig-storage] Projected configMap +STEP: Creating a kubernetes client 02/25/23 11:17:01.897 +Feb 25 11:17:01.898: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 11:17:01.9 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:17:01.924 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:17:01.933 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 +STEP: Setting up server cert 02/25/23 11:17:01.962 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 11:17:03.428 +STEP: Deploying the webhook pod 02/25/23 11:17:03.437 +STEP: Wait for the deployment to be ready 02/25/23 11:17:03.47 +Feb 25 11:17:03.492: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set +STEP: Deploying the webhook service 02/25/23 11:17:05.523 +STEP: Verifying the service has paired with the endpoint 02/25/23 11:17:05.544 +Feb 25 11:17:06.545: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] listing validating webhooks should work [Conformance] + test/e2e/apimachinery/webhook.go:581 +STEP: Listing all of the created validation webhooks 02/25/23 11:17:06.653 +STEP: Creating a configMap that does not comply to the validation webhook rules 02/25/23 11:17:06.709 +STEP: Deleting the collection of validation webhooks 02/25/23 11:17:06.764 +STEP: Creating a configMap that does not comply to the validation webhook rules 02/25/23 11:17:06.864 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:50:01.395: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-4979" for this suite. 12/14/22 16:50:01.403 -{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","completed":260,"skipped":4701,"failed":0} +Feb 25 11:17:06.911: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-4918" for this suite. 02/25/23 11:17:06.945 +STEP: Destroying namespace "webhook-4918-markers" for this suite. 02/25/23 11:17:06.963 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing validating webhooks should work [Conformance]","completed":257,"skipped":4782,"failed":0} ------------------------------ -• [SLOW TEST] [6.111 seconds] -[sig-storage] Projected configMap -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:88 +• [SLOW TEST] [5.220 seconds] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + listing validating webhooks should work [Conformance] + test/e2e/apimachinery/webhook.go:581 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected configMap + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:49:55.299 - Dec 14 16:49:55.300: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 16:49:55.302 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:49:55.328 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:49:55.332 - [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:88 - STEP: Creating configMap with name projected-configmap-test-volume-map-c8be328c-ea6d-44e8-9bbe-e5a757d39ecb 12/14/22 16:49:55.334 - STEP: Creating a pod to test consume configMaps 12/14/22 16:49:55.34 - Dec 14 16:49:55.349: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d" in namespace "projected-4979" to be "Succeeded or Failed" - Dec 14 16:49:55.353: INFO: Pod "pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d": Phase="Pending", Reason="", readiness=false. Elapsed: 3.407895ms - Dec 14 16:49:57.360: INFO: Pod "pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d": Phase="Running", Reason="", readiness=true. Elapsed: 2.010458516s - Dec 14 16:49:59.359: INFO: Pod "pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d": Phase="Running", Reason="", readiness=false. Elapsed: 4.009370593s - Dec 14 16:50:01.358: INFO: Pod "pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.008575104s - STEP: Saw pod success 12/14/22 16:50:01.358 - Dec 14 16:50:01.358: INFO: Pod "pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d" satisfied condition "Succeeded or Failed" - Dec 14 16:50:01.364: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d container agnhost-container: - STEP: delete the pod 12/14/22 16:50:01.375 - Dec 14 16:50:01.390: INFO: Waiting for pod pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d to disappear - Dec 14 16:50:01.395: INFO: Pod pod-projected-configmaps-19352ad5-5656-411d-b898-edb19849924d no longer exists - [AfterEach] [sig-storage] Projected configMap + STEP: Creating a kubernetes client 02/25/23 11:17:01.897 + Feb 25 11:17:01.898: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 11:17:01.9 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:17:01.924 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:17:01.933 + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 + STEP: Setting up server cert 02/25/23 11:17:01.962 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 11:17:03.428 + STEP: Deploying the webhook pod 02/25/23 11:17:03.437 + STEP: Wait for the deployment to be ready 02/25/23 11:17:03.47 + Feb 25 11:17:03.492: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set + STEP: Deploying the webhook service 02/25/23 11:17:05.523 + STEP: Verifying the service has paired with the endpoint 02/25/23 11:17:05.544 + Feb 25 11:17:06.545: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] listing validating webhooks should work [Conformance] + test/e2e/apimachinery/webhook.go:581 + STEP: Listing all of the created validation webhooks 02/25/23 11:17:06.653 + STEP: Creating a configMap that does not comply to the validation webhook rules 02/25/23 11:17:06.709 + STEP: Deleting the collection of validation webhooks 02/25/23 11:17:06.764 + STEP: Creating a configMap that does not comply to the validation webhook rules 02/25/23 11:17:06.864 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:50:01.395: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-4979" for this suite. 12/14/22 16:50:01.403 + Feb 25 11:17:06.911: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-4918" for this suite. 02/25/23 11:17:06.945 + STEP: Destroying namespace "webhook-4918-markers" for this suite. 02/25/23 11:17:06.963 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-network] DNS - should provide DNS for pods for Subdomain [Conformance] - test/e2e/network/dns.go:290 -[BeforeEach] [sig-network] DNS +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-network] Services + should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance] + test/e2e/network/service.go:2221 +[BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:50:01.414 -Dec 14 16:50:01.414: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename dns 12/14/22 16:50:01.416 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:50:01.442 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:50:01.446 -[It] should provide DNS for pods for Subdomain [Conformance] - test/e2e/network/dns.go:290 -STEP: Creating a test headless service 12/14/22 16:50:01.449 -STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local;sleep 1; done - 12/14/22 16:50:01.465 -STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local;sleep 1; done - 12/14/22 16:50:01.465 -STEP: creating a pod to probe DNS 12/14/22 16:50:01.465 -STEP: submitting the pod to kubernetes 12/14/22 16:50:01.466 -Dec 14 16:50:01.478: INFO: Waiting up to 15m0s for pod "dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464" in namespace "dns-4988" to be "running" -Dec 14 16:50:01.482: INFO: Pod "dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464": Phase="Pending", Reason="", readiness=false. Elapsed: 4.221656ms -Dec 14 16:50:03.492: INFO: Pod "dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01426928s -Dec 14 16:50:05.488: INFO: Pod "dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464": Phase="Running", Reason="", readiness=true. Elapsed: 4.010297831s -Dec 14 16:50:05.488: INFO: Pod "dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464" satisfied condition "running" -STEP: retrieving the pod 12/14/22 16:50:05.488 -STEP: looking for the results for each expected name from probers 12/14/22 16:50:05.496 -Dec 14 16:50:05.506: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:05.514: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:05.518: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:05.521: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:05.534: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:05.540: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:05.540: INFO: Lookups using dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local] - -Dec 14 16:50:10.559: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:10.564: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:10.579: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:10.583: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:10.583: INFO: Lookups using dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464 failed for: [wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local] - -Dec 14 16:50:15.560: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:15.566: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:15.582: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:15.587: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:15.587: INFO: Lookups using dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464 failed for: [wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local] - -Dec 14 16:50:20.560: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:20.565: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:20.581: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:20.589: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:20.589: INFO: Lookups using dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464 failed for: [wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local] - -Dec 14 16:50:25.560: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:25.564: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:25.578: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:25.583: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:25.583: INFO: Lookups using dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464 failed for: [wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local] - -Dec 14 16:50:30.570: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:30.576: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:30.591: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:30.596: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) -Dec 14 16:50:30.596: INFO: Lookups using dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464 failed for: [wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local] - -Dec 14 16:50:35.601: INFO: DNS probes using dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464 succeeded - -STEP: deleting the pod 12/14/22 16:50:35.601 -STEP: deleting the test headless service 12/14/22 16:50:35.65 -[AfterEach] [sig-network] DNS +STEP: Creating a kubernetes client 02/25/23 11:17:07.138 +Feb 25 11:17:07.138: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 11:17:07.144 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:17:07.208 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:17:07.222 +[BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 +[It] should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance] + test/e2e/network/service.go:2221 +STEP: creating service in namespace services-7643 02/25/23 11:17:07.232 +Feb 25 11:17:07.262: INFO: Waiting up to 5m0s for pod "kube-proxy-mode-detector" in namespace "services-7643" to be "running and ready" +Feb 25 11:17:07.296: INFO: Pod "kube-proxy-mode-detector": Phase="Pending", Reason="", readiness=false. Elapsed: 33.874449ms +Feb 25 11:17:07.296: INFO: The phase of Pod kube-proxy-mode-detector is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:17:09.305: INFO: Pod "kube-proxy-mode-detector": Phase="Running", Reason="", readiness=true. Elapsed: 2.042425s +Feb 25 11:17:09.305: INFO: The phase of Pod kube-proxy-mode-detector is Running (Ready = true) +Feb 25 11:17:09.305: INFO: Pod "kube-proxy-mode-detector" satisfied condition "running and ready" +Feb 25 11:17:09.311: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec kube-proxy-mode-detector -- /bin/sh -x -c curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode' +Feb 25 11:17:09.601: INFO: stderr: "+ curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode\n" +Feb 25 11:17:09.601: INFO: stdout: "iptables" +Feb 25 11:17:09.601: INFO: proxyMode: iptables +Feb 25 11:17:09.651: INFO: Waiting for pod kube-proxy-mode-detector to disappear +Feb 25 11:17:09.657: INFO: Pod kube-proxy-mode-detector no longer exists +STEP: creating service affinity-nodeport-timeout in namespace services-7643 02/25/23 11:17:09.657 +STEP: creating replication controller affinity-nodeport-timeout in namespace services-7643 02/25/23 11:17:09.682 +I0225 11:17:09.697135 13 runners.go:193] Created replication controller with name: affinity-nodeport-timeout, namespace: services-7643, replica count: 3 +I0225 11:17:12.747945 13 runners.go:193] affinity-nodeport-timeout Pods: 3 out of 3 created, 2 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0225 11:17:15.750049 13 runners.go:193] affinity-nodeport-timeout Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Feb 25 11:17:15.770: INFO: Creating new exec pod +Feb 25 11:17:15.778: INFO: Waiting up to 5m0s for pod "execpod-affinity7pqnc" in namespace "services-7643" to be "running" +Feb 25 11:17:15.784: INFO: Pod "execpod-affinity7pqnc": Phase="Pending", Reason="", readiness=false. Elapsed: 6.06376ms +Feb 25 11:17:17.791: INFO: Pod "execpod-affinity7pqnc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013071256s +Feb 25 11:17:19.792: INFO: Pod "execpod-affinity7pqnc": Phase="Running", Reason="", readiness=true. Elapsed: 4.013363605s +Feb 25 11:17:19.792: INFO: Pod "execpod-affinity7pqnc" satisfied condition "running" +Feb 25 11:17:20.799: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-nodeport-timeout 80' +Feb 25 11:17:21.092: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-nodeport-timeout 80\nConnection to affinity-nodeport-timeout 80 port [tcp/http] succeeded!\n" +Feb 25 11:17:21.092: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:17:21.092: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.28.182 80' +Feb 25 11:17:21.340: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.28.182 80\nConnection to 10.233.28.182 80 port [tcp/http] succeeded!\n" +Feb 25 11:17:21.340: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:17:21.341: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.117 31103' +Feb 25 11:17:21.557: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.117 31103\nConnection to 192.168.121.117 31103 port [tcp/*] succeeded!\n" +Feb 25 11:17:21.557: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:17:21.557: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.44 31103' +Feb 25 11:17:22.125: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.44 31103\nConnection to 192.168.121.44 31103 port [tcp/*] succeeded!\n" +Feb 25 11:17:22.125: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:17:22.127: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.250:31103/ ; done' +Feb 25 11:17:22.696: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n" +Feb 25 11:17:22.696: INFO: stdout: "\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86" +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 +Feb 25 11:17:22.697: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.250:31103/' +Feb 25 11:17:23.039: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n" +Feb 25 11:17:23.039: INFO: stdout: "affinity-nodeport-timeout-qkj86" +Feb 25 11:17:43.040: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.250:31103/' +Feb 25 11:17:43.263: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n" +Feb 25 11:17:43.263: INFO: stdout: "affinity-nodeport-timeout-qkj86" +Feb 25 11:18:03.265: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.250:31103/' +Feb 25 11:18:03.494: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n" +Feb 25 11:18:03.494: INFO: stdout: "affinity-nodeport-timeout-qkj86" +Feb 25 11:18:23.495: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.250:31103/' +Feb 25 11:18:24.066: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n" +Feb 25 11:18:24.066: INFO: stdout: "affinity-nodeport-timeout-cvkwv" +Feb 25 11:18:24.066: INFO: Cleaning up the exec pod +STEP: deleting ReplicationController affinity-nodeport-timeout in namespace services-7643, will wait for the garbage collector to delete the pods 02/25/23 11:18:24.089 +Feb 25 11:18:24.171: INFO: Deleting ReplicationController affinity-nodeport-timeout took: 25.297276ms +Feb 25 11:18:24.272: INFO: Terminating ReplicationController affinity-nodeport-timeout pods took: 101.033738ms +[AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 16:50:35.675: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "dns-4988" for this suite. 12/14/22 16:50:35.683 -{"msg":"PASSED [sig-network] DNS should provide DNS for pods for Subdomain [Conformance]","completed":261,"skipped":4751,"failed":0} +Feb 25 11:18:26.738: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-7643" for this suite. 02/25/23 11:18:26.75 +[AfterEach] [sig-network] Services + test/e2e/network/service.go:762 +{"msg":"PASSED [sig-network] Services should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance]","completed":258,"skipped":4825,"failed":0} ------------------------------ -• [SLOW TEST] [34.278 seconds] -[sig-network] DNS +• [SLOW TEST] [79.627 seconds] +[sig-network] Services test/e2e/network/common/framework.go:23 - should provide DNS for pods for Subdomain [Conformance] - test/e2e/network/dns.go:290 + should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance] + test/e2e/network/service.go:2221 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] DNS + [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:50:01.414 - Dec 14 16:50:01.414: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename dns 12/14/22 16:50:01.416 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:50:01.442 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:50:01.446 - [It] should provide DNS for pods for Subdomain [Conformance] - test/e2e/network/dns.go:290 - STEP: Creating a test headless service 12/14/22 16:50:01.449 - STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local;sleep 1; done - 12/14/22 16:50:01.465 - STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local;check="$$(dig +notcp +noall +answer +search dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service-2.dns-4988.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local;sleep 1; done - 12/14/22 16:50:01.465 - STEP: creating a pod to probe DNS 12/14/22 16:50:01.465 - STEP: submitting the pod to kubernetes 12/14/22 16:50:01.466 - Dec 14 16:50:01.478: INFO: Waiting up to 15m0s for pod "dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464" in namespace "dns-4988" to be "running" - Dec 14 16:50:01.482: INFO: Pod "dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464": Phase="Pending", Reason="", readiness=false. Elapsed: 4.221656ms - Dec 14 16:50:03.492: INFO: Pod "dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01426928s - Dec 14 16:50:05.488: INFO: Pod "dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464": Phase="Running", Reason="", readiness=true. Elapsed: 4.010297831s - Dec 14 16:50:05.488: INFO: Pod "dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464" satisfied condition "running" - STEP: retrieving the pod 12/14/22 16:50:05.488 - STEP: looking for the results for each expected name from probers 12/14/22 16:50:05.496 - Dec 14 16:50:05.506: INFO: Unable to read wheezy_udp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:05.514: INFO: Unable to read wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:05.518: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:05.521: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:05.534: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:05.540: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:05.540: INFO: Lookups using dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464 failed for: [wheezy_udp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local wheezy_tcp@dns-querier-2.dns-test-service-2.dns-4988.svc.cluster.local wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local] - - Dec 14 16:50:10.559: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:10.564: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:10.579: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:10.583: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:10.583: INFO: Lookups using dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464 failed for: [wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local] - - Dec 14 16:50:15.560: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:15.566: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:15.582: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:15.587: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:15.587: INFO: Lookups using dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464 failed for: [wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local] - - Dec 14 16:50:20.560: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:20.565: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:20.581: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:20.589: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:20.589: INFO: Lookups using dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464 failed for: [wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local] - - Dec 14 16:50:25.560: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:25.564: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:25.578: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:25.583: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:25.583: INFO: Lookups using dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464 failed for: [wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local] - - Dec 14 16:50:30.570: INFO: Unable to read wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:30.576: INFO: Unable to read wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:30.591: INFO: Unable to read jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:30.596: INFO: Unable to read jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local from pod dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464: the server could not find the requested resource (get pods dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464) - Dec 14 16:50:30.596: INFO: Lookups using dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464 failed for: [wheezy_udp@dns-test-service-2.dns-4988.svc.cluster.local wheezy_tcp@dns-test-service-2.dns-4988.svc.cluster.local jessie_udp@dns-test-service-2.dns-4988.svc.cluster.local jessie_tcp@dns-test-service-2.dns-4988.svc.cluster.local] - - Dec 14 16:50:35.601: INFO: DNS probes using dns-4988/dns-test-c84c1922-38a6-435d-a59b-b64ce5d01464 succeeded - - STEP: deleting the pod 12/14/22 16:50:35.601 - STEP: deleting the test headless service 12/14/22 16:50:35.65 - [AfterEach] [sig-network] DNS + STEP: Creating a kubernetes client 02/25/23 11:17:07.138 + Feb 25 11:17:07.138: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 11:17:07.144 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:17:07.208 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:17:07.222 + [BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 + [It] should have session affinity timeout work for NodePort service [LinuxOnly] [Conformance] + test/e2e/network/service.go:2221 + STEP: creating service in namespace services-7643 02/25/23 11:17:07.232 + Feb 25 11:17:07.262: INFO: Waiting up to 5m0s for pod "kube-proxy-mode-detector" in namespace "services-7643" to be "running and ready" + Feb 25 11:17:07.296: INFO: Pod "kube-proxy-mode-detector": Phase="Pending", Reason="", readiness=false. Elapsed: 33.874449ms + Feb 25 11:17:07.296: INFO: The phase of Pod kube-proxy-mode-detector is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:17:09.305: INFO: Pod "kube-proxy-mode-detector": Phase="Running", Reason="", readiness=true. Elapsed: 2.042425s + Feb 25 11:17:09.305: INFO: The phase of Pod kube-proxy-mode-detector is Running (Ready = true) + Feb 25 11:17:09.305: INFO: Pod "kube-proxy-mode-detector" satisfied condition "running and ready" + Feb 25 11:17:09.311: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec kube-proxy-mode-detector -- /bin/sh -x -c curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode' + Feb 25 11:17:09.601: INFO: stderr: "+ curl -q -s --connect-timeout 1 http://localhost:10249/proxyMode\n" + Feb 25 11:17:09.601: INFO: stdout: "iptables" + Feb 25 11:17:09.601: INFO: proxyMode: iptables + Feb 25 11:17:09.651: INFO: Waiting for pod kube-proxy-mode-detector to disappear + Feb 25 11:17:09.657: INFO: Pod kube-proxy-mode-detector no longer exists + STEP: creating service affinity-nodeport-timeout in namespace services-7643 02/25/23 11:17:09.657 + STEP: creating replication controller affinity-nodeport-timeout in namespace services-7643 02/25/23 11:17:09.682 + I0225 11:17:09.697135 13 runners.go:193] Created replication controller with name: affinity-nodeport-timeout, namespace: services-7643, replica count: 3 + I0225 11:17:12.747945 13 runners.go:193] affinity-nodeport-timeout Pods: 3 out of 3 created, 2 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + I0225 11:17:15.750049 13 runners.go:193] affinity-nodeport-timeout Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + Feb 25 11:17:15.770: INFO: Creating new exec pod + Feb 25 11:17:15.778: INFO: Waiting up to 5m0s for pod "execpod-affinity7pqnc" in namespace "services-7643" to be "running" + Feb 25 11:17:15.784: INFO: Pod "execpod-affinity7pqnc": Phase="Pending", Reason="", readiness=false. Elapsed: 6.06376ms + Feb 25 11:17:17.791: INFO: Pod "execpod-affinity7pqnc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013071256s + Feb 25 11:17:19.792: INFO: Pod "execpod-affinity7pqnc": Phase="Running", Reason="", readiness=true. Elapsed: 4.013363605s + Feb 25 11:17:19.792: INFO: Pod "execpod-affinity7pqnc" satisfied condition "running" + Feb 25 11:17:20.799: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-nodeport-timeout 80' + Feb 25 11:17:21.092: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-nodeport-timeout 80\nConnection to affinity-nodeport-timeout 80 port [tcp/http] succeeded!\n" + Feb 25 11:17:21.092: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:17:21.092: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.28.182 80' + Feb 25 11:17:21.340: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.28.182 80\nConnection to 10.233.28.182 80 port [tcp/http] succeeded!\n" + Feb 25 11:17:21.340: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:17:21.341: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.117 31103' + Feb 25 11:17:21.557: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.117 31103\nConnection to 192.168.121.117 31103 port [tcp/*] succeeded!\n" + Feb 25 11:17:21.557: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:17:21.557: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.44 31103' + Feb 25 11:17:22.125: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.44 31103\nConnection to 192.168.121.44 31103 port [tcp/*] succeeded!\n" + Feb 25 11:17:22.125: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:17:22.127: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.250:31103/ ; done' + Feb 25 11:17:22.696: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n" + Feb 25 11:17:22.696: INFO: stdout: "\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86\naffinity-nodeport-timeout-qkj86" + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.696: INFO: Received response from host: affinity-nodeport-timeout-qkj86 + Feb 25 11:17:22.697: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.250:31103/' + Feb 25 11:17:23.039: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n" + Feb 25 11:17:23.039: INFO: stdout: "affinity-nodeport-timeout-qkj86" + Feb 25 11:17:43.040: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.250:31103/' + Feb 25 11:17:43.263: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n" + Feb 25 11:17:43.263: INFO: stdout: "affinity-nodeport-timeout-qkj86" + Feb 25 11:18:03.265: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.250:31103/' + Feb 25 11:18:03.494: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n" + Feb 25 11:18:03.494: INFO: stdout: "affinity-nodeport-timeout-qkj86" + Feb 25 11:18:23.495: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-7643 exec execpod-affinity7pqnc -- /bin/sh -x -c curl -q -s --connect-timeout 2 http://192.168.121.250:31103/' + Feb 25 11:18:24.066: INFO: stderr: "+ curl -q -s --connect-timeout 2 http://192.168.121.250:31103/\n" + Feb 25 11:18:24.066: INFO: stdout: "affinity-nodeport-timeout-cvkwv" + Feb 25 11:18:24.066: INFO: Cleaning up the exec pod + STEP: deleting ReplicationController affinity-nodeport-timeout in namespace services-7643, will wait for the garbage collector to delete the pods 02/25/23 11:18:24.089 + Feb 25 11:18:24.171: INFO: Deleting ReplicationController affinity-nodeport-timeout took: 25.297276ms + Feb 25 11:18:24.272: INFO: Terminating ReplicationController affinity-nodeport-timeout pods took: 101.033738ms + [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 16:50:35.675: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "dns-4988" for this suite. 12/14/22 16:50:35.683 + Feb 25 11:18:26.738: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-7643" for this suite. 02/25/23 11:18:26.75 + [AfterEach] [sig-network] Services + test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Downward API volume - should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:248 -[BeforeEach] [sig-storage] Downward API volume +[sig-storage] EmptyDir volumes + should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:166 +[BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:50:35.694 -Dec 14 16:50:35.694: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 16:50:35.698 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:50:35.775 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:50:35.778 -[BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 -[It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:248 -STEP: Creating a pod to test downward API volume plugin 12/14/22 16:50:35.781 -Dec 14 16:50:35.809: INFO: Waiting up to 5m0s for pod "downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa" in namespace "downward-api-2636" to be "Succeeded or Failed" -Dec 14 16:50:35.813: INFO: Pod "downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa": Phase="Pending", Reason="", readiness=false. Elapsed: 3.848921ms -Dec 14 16:50:37.820: INFO: Pod "downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011400921s -Dec 14 16:50:39.822: INFO: Pod "downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012947662s -Dec 14 16:50:41.821: INFO: Pod "downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.012244817s -STEP: Saw pod success 12/14/22 16:50:41.821 -Dec 14 16:50:41.821: INFO: Pod "downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa" satisfied condition "Succeeded or Failed" -Dec 14 16:50:41.830: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa container client-container: -STEP: delete the pod 12/14/22 16:50:42.263 -Dec 14 16:50:42.284: INFO: Waiting for pod downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa to disappear -Dec 14 16:50:42.288: INFO: Pod downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa no longer exists -[AfterEach] [sig-storage] Downward API volume +STEP: Creating a kubernetes client 02/25/23 11:18:26.782 +Feb 25 11:18:26.782: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir 02/25/23 11:18:26.785 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:18:26.831 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:18:26.835 +[It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:166 +STEP: Creating a pod to test emptydir 0644 on node default medium 02/25/23 11:18:26.838 +Feb 25 11:18:26.852: INFO: Waiting up to 5m0s for pod "pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa" in namespace "emptydir-2385" to be "Succeeded or Failed" +Feb 25 11:18:26.862: INFO: Pod "pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa": Phase="Pending", Reason="", readiness=false. Elapsed: 10.333514ms +Feb 25 11:18:28.871: INFO: Pod "pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018808064s +Feb 25 11:18:30.873: INFO: Pod "pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa": Phase="Pending", Reason="", readiness=false. Elapsed: 4.020919056s +Feb 25 11:18:32.868: INFO: Pod "pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016707553s +STEP: Saw pod success 02/25/23 11:18:32.868 +Feb 25 11:18:32.869: INFO: Pod "pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa" satisfied condition "Succeeded or Failed" +Feb 25 11:18:32.918: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa container test-container: +STEP: delete the pod 02/25/23 11:18:32.96 +Feb 25 11:18:32.977: INFO: Waiting for pod pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa to disappear +Feb 25 11:18:32.981: INFO: Pod pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa no longer exists +[AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 -Dec 14 16:50:42.288: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-2636" for this suite. 12/14/22 16:50:42.294 -{"msg":"PASSED [sig-storage] Downward API volume should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance]","completed":262,"skipped":4778,"failed":0} +Feb 25 11:18:32.982: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-2385" for this suite. 02/25/23 11:18:32.988 +{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance]","completed":259,"skipped":4881,"failed":0} ------------------------------ -• [SLOW TEST] [6.608 seconds] -[sig-storage] Downward API volume +• [SLOW TEST] [6.218 seconds] +[sig-storage] EmptyDir volumes test/e2e/common/storage/framework.go:23 - should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:248 + should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:166 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Downward API volume + [BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:50:35.694 - Dec 14 16:50:35.694: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 16:50:35.698 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:50:35.775 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:50:35.778 - [BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 - [It] should provide node allocatable (cpu) as default cpu limit if the limit is not set [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:248 - STEP: Creating a pod to test downward API volume plugin 12/14/22 16:50:35.781 - Dec 14 16:50:35.809: INFO: Waiting up to 5m0s for pod "downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa" in namespace "downward-api-2636" to be "Succeeded or Failed" - Dec 14 16:50:35.813: INFO: Pod "downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa": Phase="Pending", Reason="", readiness=false. Elapsed: 3.848921ms - Dec 14 16:50:37.820: INFO: Pod "downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011400921s - Dec 14 16:50:39.822: INFO: Pod "downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012947662s - Dec 14 16:50:41.821: INFO: Pod "downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.012244817s - STEP: Saw pod success 12/14/22 16:50:41.821 - Dec 14 16:50:41.821: INFO: Pod "downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa" satisfied condition "Succeeded or Failed" - Dec 14 16:50:41.830: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa container client-container: - STEP: delete the pod 12/14/22 16:50:42.263 - Dec 14 16:50:42.284: INFO: Waiting for pod downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa to disappear - Dec 14 16:50:42.288: INFO: Pod downwardapi-volume-5d82689e-4b0f-470e-88aa-6576a1bb80aa no longer exists - [AfterEach] [sig-storage] Downward API volume + STEP: Creating a kubernetes client 02/25/23 11:18:26.782 + Feb 25 11:18:26.782: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir 02/25/23 11:18:26.785 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:18:26.831 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:18:26.835 + [It] should support (root,0644,default) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:166 + STEP: Creating a pod to test emptydir 0644 on node default medium 02/25/23 11:18:26.838 + Feb 25 11:18:26.852: INFO: Waiting up to 5m0s for pod "pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa" in namespace "emptydir-2385" to be "Succeeded or Failed" + Feb 25 11:18:26.862: INFO: Pod "pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa": Phase="Pending", Reason="", readiness=false. Elapsed: 10.333514ms + Feb 25 11:18:28.871: INFO: Pod "pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018808064s + Feb 25 11:18:30.873: INFO: Pod "pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa": Phase="Pending", Reason="", readiness=false. Elapsed: 4.020919056s + Feb 25 11:18:32.868: INFO: Pod "pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016707553s + STEP: Saw pod success 02/25/23 11:18:32.868 + Feb 25 11:18:32.869: INFO: Pod "pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa" satisfied condition "Succeeded or Failed" + Feb 25 11:18:32.918: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa container test-container: + STEP: delete the pod 02/25/23 11:18:32.96 + Feb 25 11:18:32.977: INFO: Waiting for pod pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa to disappear + Feb 25 11:18:32.981: INFO: Pod pod-c3d84ec2-7281-4df2-afd7-2a234f3f7caa no longer exists + [AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 - Dec 14 16:50:42.288: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-2636" for this suite. 12/14/22 16:50:42.294 + Feb 25 11:18:32.982: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-2385" for this suite. 02/25/23 11:18:32.988 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSS ------------------------------ -[sig-apps] Job - should delete a job [Conformance] - test/e2e/apps/job.go:309 -[BeforeEach] [sig-apps] Job +[sig-network] Networking Granular Checks: Pods + should function for intra-pod communication: http [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:82 +[BeforeEach] [sig-network] Networking test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:50:42.307 -Dec 14 16:50:42.307: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename job 12/14/22 16:50:42.31 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:50:42.342 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:50:42.346 -[It] should delete a job [Conformance] - test/e2e/apps/job.go:309 -STEP: Creating a job 12/14/22 16:50:42.349 -STEP: Ensuring active pods == parallelism 12/14/22 16:50:42.358 -STEP: delete a job 12/14/22 16:50:44.369 -STEP: deleting Job.batch foo in namespace job-9118, will wait for the garbage collector to delete the pods 12/14/22 16:50:44.369 -Dec 14 16:50:44.437: INFO: Deleting Job.batch foo took: 14.030519ms -Dec 14 16:50:44.538: INFO: Terminating Job.batch foo pods took: 100.815955ms -STEP: Ensuring job was deleted 12/14/22 16:51:17.639 -[AfterEach] [sig-apps] Job +STEP: Creating a kubernetes client 02/25/23 11:18:33.002 +Feb 25 11:18:33.002: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename pod-network-test 02/25/23 11:18:33.004 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:18:33.024 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:18:33.028 +[It] should function for intra-pod communication: http [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:82 +STEP: Performing setup for networking test in namespace pod-network-test-9880 02/25/23 11:18:33.033 +STEP: creating a selector 02/25/23 11:18:33.033 +STEP: Creating the service pods in kubernetes 02/25/23 11:18:33.033 +Feb 25 11:18:33.033: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable +Feb 25 11:18:33.204: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-9880" to be "running and ready" +Feb 25 11:18:33.270: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 65.678011ms +Feb 25 11:18:33.270: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:18:35.290: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 2.085883422s +Feb 25 11:18:35.290: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:18:37.275: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.071572291s +Feb 25 11:18:37.275: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:18:39.280: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.075978644s +Feb 25 11:18:39.280: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:18:41.327: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.123498024s +Feb 25 11:18:41.327: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:18:43.280: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.076354734s +Feb 25 11:18:43.280: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:18:45.278: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 12.074468293s +Feb 25 11:18:45.279: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:18:47.278: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 14.074396904s +Feb 25 11:18:47.278: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:18:49.278: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 16.074614833s +Feb 25 11:18:49.279: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:18:51.276: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 18.072190533s +Feb 25 11:18:51.276: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:18:53.280: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 20.076298794s +Feb 25 11:18:53.280: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:18:55.278: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 22.07371048s +Feb 25 11:18:55.278: INFO: The phase of Pod netserver-0 is Running (Ready = true) +Feb 25 11:18:55.278: INFO: Pod "netserver-0" satisfied condition "running and ready" +Feb 25 11:18:55.284: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-9880" to be "running and ready" +Feb 25 11:18:55.290: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 5.752018ms +Feb 25 11:18:55.290: INFO: The phase of Pod netserver-1 is Running (Ready = true) +Feb 25 11:18:55.290: INFO: Pod "netserver-1" satisfied condition "running and ready" +Feb 25 11:18:55.295: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-9880" to be "running and ready" +Feb 25 11:18:55.299: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 3.94483ms +Feb 25 11:18:55.299: INFO: The phase of Pod netserver-2 is Running (Ready = true) +Feb 25 11:18:55.299: INFO: Pod "netserver-2" satisfied condition "running and ready" +STEP: Creating test pods 02/25/23 11:18:55.304 +Feb 25 11:18:55.312: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-9880" to be "running" +Feb 25 11:18:55.321: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 9.149622ms +Feb 25 11:18:57.329: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017652151s +Feb 25 11:18:59.330: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.017878793s +Feb 25 11:18:59.330: INFO: Pod "test-container-pod" satisfied condition "running" +Feb 25 11:18:59.335: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 +Feb 25 11:18:59.335: INFO: Breadth first check of 10.233.65.21 on host 192.168.121.250... +Feb 25 11:18:59.341: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.176:9080/dial?request=hostname&protocol=http&host=10.233.65.21&port=8083&tries=1'] Namespace:pod-network-test-9880 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 11:18:59.342: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 11:18:59.343: INFO: ExecWithOptions: Clientset creation +Feb 25 11:18:59.343: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-9880/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.176%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dhttp%26host%3D10.233.65.21%26port%3D8083%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) +Feb 25 11:18:59.484: INFO: Waiting for responses: map[] +Feb 25 11:18:59.484: INFO: reached 10.233.65.21 after 0/1 tries +Feb 25 11:18:59.484: INFO: Breadth first check of 10.233.64.11 on host 192.168.121.117... +Feb 25 11:18:59.489: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.176:9080/dial?request=hostname&protocol=http&host=10.233.64.11&port=8083&tries=1'] Namespace:pod-network-test-9880 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 11:18:59.489: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 11:18:59.492: INFO: ExecWithOptions: Clientset creation +Feb 25 11:18:59.492: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-9880/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.176%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dhttp%26host%3D10.233.64.11%26port%3D8083%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) +Feb 25 11:18:59.581: INFO: Waiting for responses: map[] +Feb 25 11:18:59.581: INFO: reached 10.233.64.11 after 0/1 tries +Feb 25 11:18:59.581: INFO: Breadth first check of 10.233.66.236 on host 192.168.121.44... +Feb 25 11:18:59.587: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.176:9080/dial?request=hostname&protocol=http&host=10.233.66.236&port=8083&tries=1'] Namespace:pod-network-test-9880 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 11:18:59.587: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 11:18:59.589: INFO: ExecWithOptions: Clientset creation +Feb 25 11:18:59.589: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-9880/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.176%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dhttp%26host%3D10.233.66.236%26port%3D8083%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) +Feb 25 11:19:00.081: INFO: Waiting for responses: map[] +Feb 25 11:19:00.081: INFO: reached 10.233.66.236 after 0/1 tries +Feb 25 11:19:00.082: INFO: Going to retry 0 out of 3 pods.... +[AfterEach] [sig-network] Networking test/e2e/framework/framework.go:187 -Dec 14 16:51:17.645: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "job-9118" for this suite. 12/14/22 16:51:17.652 -{"msg":"PASSED [sig-apps] Job should delete a job [Conformance]","completed":263,"skipped":4854,"failed":0} +Feb 25 11:19:00.082: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pod-network-test-9880" for this suite. 02/25/23 11:19:00.096 +{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: http [NodeConformance] [Conformance]","completed":260,"skipped":4890,"failed":0} ------------------------------ -• [SLOW TEST] [35.355 seconds] -[sig-apps] Job -test/e2e/apps/framework.go:23 - should delete a job [Conformance] - test/e2e/apps/job.go:309 +• [SLOW TEST] [27.105 seconds] +[sig-network] Networking +test/e2e/common/network/framework.go:23 + Granular Checks: Pods + test/e2e/common/network/networking.go:32 + should function for intra-pod communication: http [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:82 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Job + [BeforeEach] [sig-network] Networking test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:50:42.307 - Dec 14 16:50:42.307: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename job 12/14/22 16:50:42.31 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:50:42.342 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:50:42.346 - [It] should delete a job [Conformance] - test/e2e/apps/job.go:309 - STEP: Creating a job 12/14/22 16:50:42.349 - STEP: Ensuring active pods == parallelism 12/14/22 16:50:42.358 - STEP: delete a job 12/14/22 16:50:44.369 - STEP: deleting Job.batch foo in namespace job-9118, will wait for the garbage collector to delete the pods 12/14/22 16:50:44.369 - Dec 14 16:50:44.437: INFO: Deleting Job.batch foo took: 14.030519ms - Dec 14 16:50:44.538: INFO: Terminating Job.batch foo pods took: 100.815955ms - STEP: Ensuring job was deleted 12/14/22 16:51:17.639 - [AfterEach] [sig-apps] Job + STEP: Creating a kubernetes client 02/25/23 11:18:33.002 + Feb 25 11:18:33.002: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename pod-network-test 02/25/23 11:18:33.004 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:18:33.024 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:18:33.028 + [It] should function for intra-pod communication: http [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:82 + STEP: Performing setup for networking test in namespace pod-network-test-9880 02/25/23 11:18:33.033 + STEP: creating a selector 02/25/23 11:18:33.033 + STEP: Creating the service pods in kubernetes 02/25/23 11:18:33.033 + Feb 25 11:18:33.033: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable + Feb 25 11:18:33.204: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-9880" to be "running and ready" + Feb 25 11:18:33.270: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 65.678011ms + Feb 25 11:18:33.270: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:18:35.290: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 2.085883422s + Feb 25 11:18:35.290: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:18:37.275: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.071572291s + Feb 25 11:18:37.275: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:18:39.280: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.075978644s + Feb 25 11:18:39.280: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:18:41.327: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.123498024s + Feb 25 11:18:41.327: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:18:43.280: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.076354734s + Feb 25 11:18:43.280: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:18:45.278: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 12.074468293s + Feb 25 11:18:45.279: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:18:47.278: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 14.074396904s + Feb 25 11:18:47.278: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:18:49.278: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 16.074614833s + Feb 25 11:18:49.279: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:18:51.276: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 18.072190533s + Feb 25 11:18:51.276: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:18:53.280: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 20.076298794s + Feb 25 11:18:53.280: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:18:55.278: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 22.07371048s + Feb 25 11:18:55.278: INFO: The phase of Pod netserver-0 is Running (Ready = true) + Feb 25 11:18:55.278: INFO: Pod "netserver-0" satisfied condition "running and ready" + Feb 25 11:18:55.284: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-9880" to be "running and ready" + Feb 25 11:18:55.290: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 5.752018ms + Feb 25 11:18:55.290: INFO: The phase of Pod netserver-1 is Running (Ready = true) + Feb 25 11:18:55.290: INFO: Pod "netserver-1" satisfied condition "running and ready" + Feb 25 11:18:55.295: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-9880" to be "running and ready" + Feb 25 11:18:55.299: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 3.94483ms + Feb 25 11:18:55.299: INFO: The phase of Pod netserver-2 is Running (Ready = true) + Feb 25 11:18:55.299: INFO: Pod "netserver-2" satisfied condition "running and ready" + STEP: Creating test pods 02/25/23 11:18:55.304 + Feb 25 11:18:55.312: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-9880" to be "running" + Feb 25 11:18:55.321: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 9.149622ms + Feb 25 11:18:57.329: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017652151s + Feb 25 11:18:59.330: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.017878793s + Feb 25 11:18:59.330: INFO: Pod "test-container-pod" satisfied condition "running" + Feb 25 11:18:59.335: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 + Feb 25 11:18:59.335: INFO: Breadth first check of 10.233.65.21 on host 192.168.121.250... + Feb 25 11:18:59.341: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.176:9080/dial?request=hostname&protocol=http&host=10.233.65.21&port=8083&tries=1'] Namespace:pod-network-test-9880 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 11:18:59.342: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 11:18:59.343: INFO: ExecWithOptions: Clientset creation + Feb 25 11:18:59.343: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-9880/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.176%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dhttp%26host%3D10.233.65.21%26port%3D8083%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) + Feb 25 11:18:59.484: INFO: Waiting for responses: map[] + Feb 25 11:18:59.484: INFO: reached 10.233.65.21 after 0/1 tries + Feb 25 11:18:59.484: INFO: Breadth first check of 10.233.64.11 on host 192.168.121.117... + Feb 25 11:18:59.489: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.176:9080/dial?request=hostname&protocol=http&host=10.233.64.11&port=8083&tries=1'] Namespace:pod-network-test-9880 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 11:18:59.489: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 11:18:59.492: INFO: ExecWithOptions: Clientset creation + Feb 25 11:18:59.492: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-9880/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.176%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dhttp%26host%3D10.233.64.11%26port%3D8083%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) + Feb 25 11:18:59.581: INFO: Waiting for responses: map[] + Feb 25 11:18:59.581: INFO: reached 10.233.64.11 after 0/1 tries + Feb 25 11:18:59.581: INFO: Breadth first check of 10.233.66.236 on host 192.168.121.44... + Feb 25 11:18:59.587: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.176:9080/dial?request=hostname&protocol=http&host=10.233.66.236&port=8083&tries=1'] Namespace:pod-network-test-9880 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 11:18:59.587: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 11:18:59.589: INFO: ExecWithOptions: Clientset creation + Feb 25 11:18:59.589: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-9880/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.176%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dhttp%26host%3D10.233.66.236%26port%3D8083%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) + Feb 25 11:19:00.081: INFO: Waiting for responses: map[] + Feb 25 11:19:00.081: INFO: reached 10.233.66.236 after 0/1 tries + Feb 25 11:19:00.082: INFO: Going to retry 0 out of 3 pods.... + [AfterEach] [sig-network] Networking test/e2e/framework/framework.go:187 - Dec 14 16:51:17.645: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "job-9118" for this suite. 12/14/22 16:51:17.652 + Feb 25 11:19:00.082: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "pod-network-test-9880" for this suite. 02/25/23 11:19:00.096 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] - should be able to convert a non homogeneous list of CRs [Conformance] - test/e2e/apimachinery/crd_conversion_webhook.go:184 -[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] +[sig-node] RuntimeClass + should schedule a Pod requesting a RuntimeClass without PodOverhead [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:104 +[BeforeEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:51:17.663 -Dec 14 16:51:17.663: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename crd-webhook 12/14/22 16:51:17.665 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:51:17.686 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:51:17.689 -[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/crd_conversion_webhook.go:128 -STEP: Setting up server cert 12/14/22 16:51:17.693 -STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication 12/14/22 16:51:19.33 -STEP: Deploying the custom resource conversion webhook pod 12/14/22 16:51:19.343 -STEP: Wait for the deployment to be ready 12/14/22 16:51:19.359 -Dec 14 16:51:19.374: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set -Dec 14 16:51:21.402: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 51, 19, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 51, 19, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 51, 19, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 51, 19, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-59dfc5db8d\" is progressing."}}, CollisionCount:(*int32)(nil)} -STEP: Deploying the webhook service 12/14/22 16:51:23.41 -STEP: Verifying the service has paired with the endpoint 12/14/22 16:51:23.429 -Dec 14 16:51:24.430: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1 -[It] should be able to convert a non homogeneous list of CRs [Conformance] - test/e2e/apimachinery/crd_conversion_webhook.go:184 -Dec 14 16:51:24.437: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Creating a v1 custom resource 12/14/22 16:51:27.206 -STEP: Create a v2 custom resource 12/14/22 16:51:27.234 -STEP: List CRs in v1 12/14/22 16:51:27.398 -STEP: List CRs in v2 12/14/22 16:51:27.408 -[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 11:19:00.114 +Feb 25 11:19:00.114: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename runtimeclass 02/25/23 11:19:00.117 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:19:00.148 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:19:00.152 +[It] should schedule a Pod requesting a RuntimeClass without PodOverhead [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:104 +Feb 25 11:19:00.174: INFO: Waiting up to 1m20s for at least 1 pods in namespace runtimeclass-9222 to be scheduled +Feb 25 11:19:00.183: INFO: 1 pods are not scheduled: [runtimeclass-9222/test-runtimeclass-runtimeclass-9222-preconfigured-handler-2lcnb(7e9462f4-1bb2-495a-a874-ddc1b8cc614e)] +[AfterEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:187 -Dec 14 16:51:27.930: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-webhook-7946" for this suite. 12/14/22 16:51:27.937 -[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/crd_conversion_webhook.go:139 -{"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert a non homogeneous list of CRs [Conformance]","completed":264,"skipped":4857,"failed":0} +Feb 25 11:19:02.209: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "runtimeclass-9222" for this suite. 02/25/23 11:19:02.224 +{"msg":"PASSED [sig-node] RuntimeClass should schedule a Pod requesting a RuntimeClass without PodOverhead [NodeConformance] [Conformance]","completed":261,"skipped":4920,"failed":0} ------------------------------ -• [SLOW TEST] [10.443 seconds] -[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - should be able to convert a non homogeneous list of CRs [Conformance] - test/e2e/apimachinery/crd_conversion_webhook.go:184 +• [2.125 seconds] +[sig-node] RuntimeClass +test/e2e/common/node/framework.go:23 + should schedule a Pod requesting a RuntimeClass without PodOverhead [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:104 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] + [BeforeEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:51:17.663 - Dec 14 16:51:17.663: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename crd-webhook 12/14/22 16:51:17.665 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:51:17.686 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:51:17.689 - [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/crd_conversion_webhook.go:128 - STEP: Setting up server cert 12/14/22 16:51:17.693 - STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication 12/14/22 16:51:19.33 - STEP: Deploying the custom resource conversion webhook pod 12/14/22 16:51:19.343 - STEP: Wait for the deployment to be ready 12/14/22 16:51:19.359 - Dec 14 16:51:19.374: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set - Dec 14 16:51:21.402: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 51, 19, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 51, 19, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 51, 19, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 51, 19, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-59dfc5db8d\" is progressing."}}, CollisionCount:(*int32)(nil)} - STEP: Deploying the webhook service 12/14/22 16:51:23.41 - STEP: Verifying the service has paired with the endpoint 12/14/22 16:51:23.429 - Dec 14 16:51:24.430: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1 - [It] should be able to convert a non homogeneous list of CRs [Conformance] - test/e2e/apimachinery/crd_conversion_webhook.go:184 - Dec 14 16:51:24.437: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Creating a v1 custom resource 12/14/22 16:51:27.206 - STEP: Create a v2 custom resource 12/14/22 16:51:27.234 - STEP: List CRs in v1 12/14/22 16:51:27.398 - STEP: List CRs in v2 12/14/22 16:51:27.408 - [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 11:19:00.114 + Feb 25 11:19:00.114: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename runtimeclass 02/25/23 11:19:00.117 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:19:00.148 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:19:00.152 + [It] should schedule a Pod requesting a RuntimeClass without PodOverhead [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:104 + Feb 25 11:19:00.174: INFO: Waiting up to 1m20s for at least 1 pods in namespace runtimeclass-9222 to be scheduled + Feb 25 11:19:00.183: INFO: 1 pods are not scheduled: [runtimeclass-9222/test-runtimeclass-runtimeclass-9222-preconfigured-handler-2lcnb(7e9462f4-1bb2-495a-a874-ddc1b8cc614e)] + [AfterEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:187 - Dec 14 16:51:27.930: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "crd-webhook-7946" for this suite. 12/14/22 16:51:27.937 - [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/crd_conversion_webhook.go:139 + Feb 25 11:19:02.209: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "runtimeclass-9222" for this suite. 02/25/23 11:19:02.224 << End Captured GinkgoWriter Output ------------------------------ -S +SSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-cli] Kubectl client Kubectl diff - should check if kubectl diff finds a difference for Deployments [Conformance] - test/e2e/kubectl/kubectl.go:929 -[BeforeEach] [sig-cli] Kubectl client +[sig-storage] ConfigMap + should be consumable from pods in volume as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:73 +[BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:51:28.111 -Dec 14 16:51:28.112: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 16:51:28.116 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:51:28.221 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:51:28.231 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 -[It] should check if kubectl diff finds a difference for Deployments [Conformance] - test/e2e/kubectl/kubectl.go:929 -STEP: create deployment with httpd image 12/14/22 16:51:28.238 -Dec 14 16:51:28.241: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-9718 create -f -' -Dec 14 16:51:29.724: INFO: stderr: "" -Dec 14 16:51:29.724: INFO: stdout: "deployment.apps/httpd-deployment created\n" -STEP: verify diff finds difference between live and declared image 12/14/22 16:51:29.724 -Dec 14 16:51:29.726: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-9718 diff -f -' -Dec 14 16:51:31.160: INFO: rc: 1 -Dec 14 16:51:31.161: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-9718 delete -f -' -Dec 14 16:51:31.339: INFO: stderr: "" -Dec 14 16:51:31.339: INFO: stdout: "deployment.apps \"httpd-deployment\" deleted\n" -[AfterEach] [sig-cli] Kubectl client +STEP: Creating a kubernetes client 02/25/23 11:19:02.249 +Feb 25 11:19:02.249: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename configmap 02/25/23 11:19:02.253 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:19:02.331 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:19:02.339 +[It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:73 +STEP: Creating configMap with name configmap-test-volume-321ba5a8-7b1e-4aa4-8641-3c475570b0f8 02/25/23 11:19:02.352 +STEP: Creating a pod to test consume configMaps 02/25/23 11:19:02.36 +Feb 25 11:19:02.375: INFO: Waiting up to 5m0s for pod "pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef" in namespace "configmap-9346" to be "Succeeded or Failed" +Feb 25 11:19:02.384: INFO: Pod "pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef": Phase="Pending", Reason="", readiness=false. Elapsed: 9.223013ms +Feb 25 11:19:04.391: INFO: Pod "pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01636589s +Feb 25 11:19:06.402: INFO: Pod "pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef": Phase="Pending", Reason="", readiness=false. Elapsed: 4.026964704s +Feb 25 11:19:08.392: INFO: Pod "pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016860071s +STEP: Saw pod success 02/25/23 11:19:08.392 +Feb 25 11:19:08.392: INFO: Pod "pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef" satisfied condition "Succeeded or Failed" +Feb 25 11:19:08.399: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef container agnhost-container: +STEP: delete the pod 02/25/23 11:19:08.411 +Feb 25 11:19:08.429: INFO: Waiting for pod pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef to disappear +Feb 25 11:19:08.436: INFO: Pod pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef no longer exists +[AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 -Dec 14 16:51:31.339: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-9718" for this suite. 12/14/22 16:51:31.348 -{"msg":"PASSED [sig-cli] Kubectl client Kubectl diff should check if kubectl diff finds a difference for Deployments [Conformance]","completed":265,"skipped":4858,"failed":0} +Feb 25 11:19:08.436: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-9346" for this suite. 02/25/23 11:19:08.444 +{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]","completed":262,"skipped":4939,"failed":0} ------------------------------ -• [3.246 seconds] -[sig-cli] Kubectl client -test/e2e/kubectl/framework.go:23 - Kubectl diff - test/e2e/kubectl/kubectl.go:923 - should check if kubectl diff finds a difference for Deployments [Conformance] - test/e2e/kubectl/kubectl.go:929 +• [SLOW TEST] [6.208 seconds] +[sig-storage] ConfigMap +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:73 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-cli] Kubectl client + [BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:51:28.111 - Dec 14 16:51:28.112: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 16:51:28.116 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:51:28.221 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:51:28.231 - [BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 - [It] should check if kubectl diff finds a difference for Deployments [Conformance] - test/e2e/kubectl/kubectl.go:929 - STEP: create deployment with httpd image 12/14/22 16:51:28.238 - Dec 14 16:51:28.241: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-9718 create -f -' - Dec 14 16:51:29.724: INFO: stderr: "" - Dec 14 16:51:29.724: INFO: stdout: "deployment.apps/httpd-deployment created\n" - STEP: verify diff finds difference between live and declared image 12/14/22 16:51:29.724 - Dec 14 16:51:29.726: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-9718 diff -f -' - Dec 14 16:51:31.160: INFO: rc: 1 - Dec 14 16:51:31.161: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-9718 delete -f -' - Dec 14 16:51:31.339: INFO: stderr: "" - Dec 14 16:51:31.339: INFO: stdout: "deployment.apps \"httpd-deployment\" deleted\n" - [AfterEach] [sig-cli] Kubectl client + STEP: Creating a kubernetes client 02/25/23 11:19:02.249 + Feb 25 11:19:02.249: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename configmap 02/25/23 11:19:02.253 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:19:02.331 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:19:02.339 + [It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:73 + STEP: Creating configMap with name configmap-test-volume-321ba5a8-7b1e-4aa4-8641-3c475570b0f8 02/25/23 11:19:02.352 + STEP: Creating a pod to test consume configMaps 02/25/23 11:19:02.36 + Feb 25 11:19:02.375: INFO: Waiting up to 5m0s for pod "pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef" in namespace "configmap-9346" to be "Succeeded or Failed" + Feb 25 11:19:02.384: INFO: Pod "pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef": Phase="Pending", Reason="", readiness=false. Elapsed: 9.223013ms + Feb 25 11:19:04.391: INFO: Pod "pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01636589s + Feb 25 11:19:06.402: INFO: Pod "pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef": Phase="Pending", Reason="", readiness=false. Elapsed: 4.026964704s + Feb 25 11:19:08.392: INFO: Pod "pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016860071s + STEP: Saw pod success 02/25/23 11:19:08.392 + Feb 25 11:19:08.392: INFO: Pod "pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef" satisfied condition "Succeeded or Failed" + Feb 25 11:19:08.399: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef container agnhost-container: + STEP: delete the pod 02/25/23 11:19:08.411 + Feb 25 11:19:08.429: INFO: Waiting for pod pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef to disappear + Feb 25 11:19:08.436: INFO: Pod pod-configmaps-6b4e7b1e-8391-498a-aa2f-6ee2ceb0dbef no longer exists + [AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 - Dec 14 16:51:31.339: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-9718" for this suite. 12/14/22 16:51:31.348 + Feb 25 11:19:08.436: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "configmap-9346" for this suite. 02/25/23 11:19:08.444 << End Captured GinkgoWriter Output ------------------------------ -[sig-node] Pods - should support retrieving logs from the container over websockets [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:617 -[BeforeEach] [sig-node] Pods +S +------------------------------ +[sig-auth] ServiceAccounts + ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance] + test/e2e/auth/service_accounts.go:528 +[BeforeEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:51:31.357 -Dec 14 16:51:31.358: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename pods 12/14/22 16:51:31.361 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:51:31.39 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:51:31.393 -[BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 -[It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:617 -Dec 14 16:51:31.396: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: creating the pod 12/14/22 16:51:31.397 -STEP: submitting the pod to kubernetes 12/14/22 16:51:31.397 -Dec 14 16:51:31.408: INFO: Waiting up to 5m0s for pod "pod-logs-websocket-1a60324e-e521-4999-8660-e29787d531a4" in namespace "pods-7737" to be "running and ready" -Dec 14 16:51:31.412: INFO: Pod "pod-logs-websocket-1a60324e-e521-4999-8660-e29787d531a4": Phase="Pending", Reason="", readiness=false. Elapsed: 3.677487ms -Dec 14 16:51:31.412: INFO: The phase of Pod pod-logs-websocket-1a60324e-e521-4999-8660-e29787d531a4 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:51:33.419: INFO: Pod "pod-logs-websocket-1a60324e-e521-4999-8660-e29787d531a4": Phase="Running", Reason="", readiness=true. Elapsed: 2.010664715s -Dec 14 16:51:33.419: INFO: The phase of Pod pod-logs-websocket-1a60324e-e521-4999-8660-e29787d531a4 is Running (Ready = true) -Dec 14 16:51:33.419: INFO: Pod "pod-logs-websocket-1a60324e-e521-4999-8660-e29787d531a4" satisfied condition "running and ready" -[AfterEach] [sig-node] Pods +STEP: Creating a kubernetes client 02/25/23 11:19:08.457 +Feb 25 11:19:08.457: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename svcaccounts 02/25/23 11:19:08.463 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:19:08.487 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:19:08.494 +[It] ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance] + test/e2e/auth/service_accounts.go:528 +Feb 25 11:19:08.518: INFO: created pod +Feb 25 11:19:08.518: INFO: Waiting up to 5m0s for pod "oidc-discovery-validator" in namespace "svcaccounts-7088" to be "Succeeded or Failed" +Feb 25 11:19:08.522: INFO: Pod "oidc-discovery-validator": Phase="Pending", Reason="", readiness=false. Elapsed: 3.601438ms +Feb 25 11:19:10.529: INFO: Pod "oidc-discovery-validator": Phase="Running", Reason="", readiness=true. Elapsed: 2.010774642s +Feb 25 11:19:12.535: INFO: Pod "oidc-discovery-validator": Phase="Running", Reason="", readiness=false. Elapsed: 4.016335529s +Feb 25 11:19:14.529: INFO: Pod "oidc-discovery-validator": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.010550244s +STEP: Saw pod success 02/25/23 11:19:14.529 +Feb 25 11:19:14.529: INFO: Pod "oidc-discovery-validator" satisfied condition "Succeeded or Failed" +Feb 25 11:19:44.531: INFO: polling logs +Feb 25 11:19:44.544: INFO: Pod logs: +I0225 11:19:09.543270 1 log.go:195] OK: Got token +I0225 11:19:09.543601 1 log.go:195] validating with in-cluster discovery +I0225 11:19:09.550981 1 log.go:195] OK: got issuer https://kubernetes.default.svc.cluster.local +I0225 11:19:09.552162 1 log.go:195] Full, not-validated claims: +openidmetadata.claims{Claims:jwt.Claims{Issuer:"https://kubernetes.default.svc.cluster.local", Subject:"system:serviceaccount:svcaccounts-7088:default", Audience:jwt.Audience{"oidc-discovery-test"}, Expiry:1677324548, NotBefore:1677323948, IssuedAt:1677323948, ID:""}, Kubernetes:openidmetadata.kubeClaims{Namespace:"svcaccounts-7088", ServiceAccount:openidmetadata.kubeName{Name:"default", UID:"1d592386-cc72-4a3c-97de-d7c760e7a69f"}}} +I0225 11:19:09.582035 1 log.go:195] OK: Constructed OIDC provider for issuer https://kubernetes.default.svc.cluster.local +I0225 11:19:09.591408 1 log.go:195] OK: Validated signature on JWT +I0225 11:19:09.591611 1 log.go:195] OK: Got valid claims from token! +I0225 11:19:09.591706 1 log.go:195] Full, validated claims: +&openidmetadata.claims{Claims:jwt.Claims{Issuer:"https://kubernetes.default.svc.cluster.local", Subject:"system:serviceaccount:svcaccounts-7088:default", Audience:jwt.Audience{"oidc-discovery-test"}, Expiry:1677324548, NotBefore:1677323948, IssuedAt:1677323948, ID:""}, Kubernetes:openidmetadata.kubeClaims{Namespace:"svcaccounts-7088", ServiceAccount:openidmetadata.kubeName{Name:"default", UID:"1d592386-cc72-4a3c-97de-d7c760e7a69f"}}} + +Feb 25 11:19:44.544: INFO: completed pod +[AfterEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:187 -Dec 14 16:51:33.456: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-7737" for this suite. 12/14/22 16:51:33.462 -{"msg":"PASSED [sig-node] Pods should support retrieving logs from the container over websockets [NodeConformance] [Conformance]","completed":266,"skipped":4858,"failed":0} +Feb 25 11:19:44.554: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "svcaccounts-7088" for this suite. 02/25/23 11:19:44.563 +{"msg":"PASSED [sig-auth] ServiceAccounts ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance]","completed":263,"skipped":4940,"failed":0} ------------------------------ -• [2.116 seconds] -[sig-node] Pods -test/e2e/common/node/framework.go:23 - should support retrieving logs from the container over websockets [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:617 +• [SLOW TEST] [36.118 seconds] +[sig-auth] ServiceAccounts +test/e2e/auth/framework.go:23 + ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance] + test/e2e/auth/service_accounts.go:528 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Pods + [BeforeEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:51:31.357 - Dec 14 16:51:31.358: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename pods 12/14/22 16:51:31.361 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:51:31.39 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:51:31.393 - [BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 - [It] should support retrieving logs from the container over websockets [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:617 - Dec 14 16:51:31.396: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: creating the pod 12/14/22 16:51:31.397 - STEP: submitting the pod to kubernetes 12/14/22 16:51:31.397 - Dec 14 16:51:31.408: INFO: Waiting up to 5m0s for pod "pod-logs-websocket-1a60324e-e521-4999-8660-e29787d531a4" in namespace "pods-7737" to be "running and ready" - Dec 14 16:51:31.412: INFO: Pod "pod-logs-websocket-1a60324e-e521-4999-8660-e29787d531a4": Phase="Pending", Reason="", readiness=false. Elapsed: 3.677487ms - Dec 14 16:51:31.412: INFO: The phase of Pod pod-logs-websocket-1a60324e-e521-4999-8660-e29787d531a4 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:51:33.419: INFO: Pod "pod-logs-websocket-1a60324e-e521-4999-8660-e29787d531a4": Phase="Running", Reason="", readiness=true. Elapsed: 2.010664715s - Dec 14 16:51:33.419: INFO: The phase of Pod pod-logs-websocket-1a60324e-e521-4999-8660-e29787d531a4 is Running (Ready = true) - Dec 14 16:51:33.419: INFO: Pod "pod-logs-websocket-1a60324e-e521-4999-8660-e29787d531a4" satisfied condition "running and ready" - [AfterEach] [sig-node] Pods + STEP: Creating a kubernetes client 02/25/23 11:19:08.457 + Feb 25 11:19:08.457: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename svcaccounts 02/25/23 11:19:08.463 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:19:08.487 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:19:08.494 + [It] ServiceAccountIssuerDiscovery should support OIDC discovery of service account issuer [Conformance] + test/e2e/auth/service_accounts.go:528 + Feb 25 11:19:08.518: INFO: created pod + Feb 25 11:19:08.518: INFO: Waiting up to 5m0s for pod "oidc-discovery-validator" in namespace "svcaccounts-7088" to be "Succeeded or Failed" + Feb 25 11:19:08.522: INFO: Pod "oidc-discovery-validator": Phase="Pending", Reason="", readiness=false. Elapsed: 3.601438ms + Feb 25 11:19:10.529: INFO: Pod "oidc-discovery-validator": Phase="Running", Reason="", readiness=true. Elapsed: 2.010774642s + Feb 25 11:19:12.535: INFO: Pod "oidc-discovery-validator": Phase="Running", Reason="", readiness=false. Elapsed: 4.016335529s + Feb 25 11:19:14.529: INFO: Pod "oidc-discovery-validator": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.010550244s + STEP: Saw pod success 02/25/23 11:19:14.529 + Feb 25 11:19:14.529: INFO: Pod "oidc-discovery-validator" satisfied condition "Succeeded or Failed" + Feb 25 11:19:44.531: INFO: polling logs + Feb 25 11:19:44.544: INFO: Pod logs: + I0225 11:19:09.543270 1 log.go:195] OK: Got token + I0225 11:19:09.543601 1 log.go:195] validating with in-cluster discovery + I0225 11:19:09.550981 1 log.go:195] OK: got issuer https://kubernetes.default.svc.cluster.local + I0225 11:19:09.552162 1 log.go:195] Full, not-validated claims: + openidmetadata.claims{Claims:jwt.Claims{Issuer:"https://kubernetes.default.svc.cluster.local", Subject:"system:serviceaccount:svcaccounts-7088:default", Audience:jwt.Audience{"oidc-discovery-test"}, Expiry:1677324548, NotBefore:1677323948, IssuedAt:1677323948, ID:""}, Kubernetes:openidmetadata.kubeClaims{Namespace:"svcaccounts-7088", ServiceAccount:openidmetadata.kubeName{Name:"default", UID:"1d592386-cc72-4a3c-97de-d7c760e7a69f"}}} + I0225 11:19:09.582035 1 log.go:195] OK: Constructed OIDC provider for issuer https://kubernetes.default.svc.cluster.local + I0225 11:19:09.591408 1 log.go:195] OK: Validated signature on JWT + I0225 11:19:09.591611 1 log.go:195] OK: Got valid claims from token! + I0225 11:19:09.591706 1 log.go:195] Full, validated claims: + &openidmetadata.claims{Claims:jwt.Claims{Issuer:"https://kubernetes.default.svc.cluster.local", Subject:"system:serviceaccount:svcaccounts-7088:default", Audience:jwt.Audience{"oidc-discovery-test"}, Expiry:1677324548, NotBefore:1677323948, IssuedAt:1677323948, ID:""}, Kubernetes:openidmetadata.kubeClaims{Namespace:"svcaccounts-7088", ServiceAccount:openidmetadata.kubeName{Name:"default", UID:"1d592386-cc72-4a3c-97de-d7c760e7a69f"}}} + + Feb 25 11:19:44.544: INFO: completed pod + [AfterEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:187 - Dec 14 16:51:33.456: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "pods-7737" for this suite. 12/14/22 16:51:33.462 + Feb 25 11:19:44.554: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "svcaccounts-7088" for this suite. 02/25/23 11:19:44.563 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSS +SSSSSS ------------------------------ -[sig-node] NoExecuteTaintManager Multiple Pods [Serial] - evicts pods with minTolerationSeconds [Disruptive] [Conformance] - test/e2e/node/taints.go:420 -[BeforeEach] [sig-node] NoExecuteTaintManager Multiple Pods [Serial] +[sig-storage] Projected secret + should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:45 +[BeforeEach] [sig-storage] Projected secret test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:51:33.477 -Dec 14 16:51:33.477: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename taint-multiple-pods 12/14/22 16:51:33.478 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:51:33.509 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:51:33.512 -[BeforeEach] [sig-node] NoExecuteTaintManager Multiple Pods [Serial] - test/e2e/node/taints.go:348 -Dec 14 16:51:33.516: INFO: Waiting up to 1m0s for all nodes to be ready -Dec 14 16:52:33.545: INFO: Waiting for terminating namespaces to be deleted... -[It] evicts pods with minTolerationSeconds [Disruptive] [Conformance] - test/e2e/node/taints.go:420 -Dec 14 16:52:33.550: INFO: Starting informer... -STEP: Starting pods... 12/14/22 16:52:33.55 -Dec 14 16:52:33.788: INFO: Pod1 is running on kiedai4iakae-3. Tainting Node -Dec 14 16:52:34.004: INFO: Waiting up to 5m0s for pod "taint-eviction-b1" in namespace "taint-multiple-pods-5248" to be "running" -Dec 14 16:52:34.012: INFO: Pod "taint-eviction-b1": Phase="Pending", Reason="", readiness=false. Elapsed: 7.941817ms -Dec 14 16:52:36.019: INFO: Pod "taint-eviction-b1": Phase="Running", Reason="", readiness=true. Elapsed: 2.014613942s -Dec 14 16:52:36.019: INFO: Pod "taint-eviction-b1" satisfied condition "running" -Dec 14 16:52:36.019: INFO: Waiting up to 5m0s for pod "taint-eviction-b2" in namespace "taint-multiple-pods-5248" to be "running" -Dec 14 16:52:36.023: INFO: Pod "taint-eviction-b2": Phase="Running", Reason="", readiness=true. Elapsed: 4.660032ms -Dec 14 16:52:36.023: INFO: Pod "taint-eviction-b2" satisfied condition "running" -Dec 14 16:52:36.024: INFO: Pod2 is running on kiedai4iakae-3. Tainting Node -STEP: Trying to apply a taint on the Node 12/14/22 16:52:36.024 -STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 12/14/22 16:52:36.042 -STEP: Waiting for Pod1 and Pod2 to be deleted 12/14/22 16:52:36.048 -Dec 14 16:52:41.592: INFO: Noticed Pod "taint-eviction-b1" gets evicted. -Dec 14 16:53:01.632: INFO: Noticed Pod "taint-eviction-b2" gets evicted. -STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 12/14/22 16:53:01.65 -[AfterEach] [sig-node] NoExecuteTaintManager Multiple Pods [Serial] +STEP: Creating a kubernetes client 02/25/23 11:19:44.576 +Feb 25 11:19:44.576: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 11:19:44.581 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:19:44.617 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:19:44.632 +[It] should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:45 +STEP: Creating projection with secret that has name projected-secret-test-1efe4611-3488-4a7b-8f96-2296eb408b61 02/25/23 11:19:44.645 +STEP: Creating a pod to test consume secrets 02/25/23 11:19:44.668 +Feb 25 11:19:44.698: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302" in namespace "projected-8587" to be "Succeeded or Failed" +Feb 25 11:19:44.709: INFO: Pod "pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302": Phase="Pending", Reason="", readiness=false. Elapsed: 10.599493ms +Feb 25 11:19:46.717: INFO: Pod "pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018960503s +Feb 25 11:19:48.716: INFO: Pod "pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302": Phase="Pending", Reason="", readiness=false. Elapsed: 4.017035685s +Feb 25 11:19:50.717: INFO: Pod "pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.018719967s +STEP: Saw pod success 02/25/23 11:19:50.718 +Feb 25 11:19:50.718: INFO: Pod "pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302" satisfied condition "Succeeded or Failed" +Feb 25 11:19:50.723: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302 container projected-secret-volume-test: +STEP: delete the pod 02/25/23 11:19:50.736 +Feb 25 11:19:50.761: INFO: Waiting for pod pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302 to disappear +Feb 25 11:19:50.767: INFO: Pod pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302 no longer exists +[AfterEach] [sig-storage] Projected secret test/e2e/framework/framework.go:187 -Dec 14 16:53:01.661: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "taint-multiple-pods-5248" for this suite. 12/14/22 16:53:01.668 -{"msg":"PASSED [sig-node] NoExecuteTaintManager Multiple Pods [Serial] evicts pods with minTolerationSeconds [Disruptive] [Conformance]","completed":267,"skipped":4872,"failed":0} +Feb 25 11:19:50.767: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-8587" for this suite. 02/25/23 11:19:50.773 +{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume [NodeConformance] [Conformance]","completed":264,"skipped":4946,"failed":0} ------------------------------ -• [SLOW TEST] [88.204 seconds] -[sig-node] NoExecuteTaintManager Multiple Pods [Serial] -test/e2e/node/framework.go:23 - evicts pods with minTolerationSeconds [Disruptive] [Conformance] - test/e2e/node/taints.go:420 +• [SLOW TEST] [6.208 seconds] +[sig-storage] Projected secret +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:45 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] NoExecuteTaintManager Multiple Pods [Serial] + [BeforeEach] [sig-storage] Projected secret test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:51:33.477 - Dec 14 16:51:33.477: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename taint-multiple-pods 12/14/22 16:51:33.478 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:51:33.509 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:51:33.512 - [BeforeEach] [sig-node] NoExecuteTaintManager Multiple Pods [Serial] - test/e2e/node/taints.go:348 - Dec 14 16:51:33.516: INFO: Waiting up to 1m0s for all nodes to be ready - Dec 14 16:52:33.545: INFO: Waiting for terminating namespaces to be deleted... - [It] evicts pods with minTolerationSeconds [Disruptive] [Conformance] - test/e2e/node/taints.go:420 - Dec 14 16:52:33.550: INFO: Starting informer... - STEP: Starting pods... 12/14/22 16:52:33.55 - Dec 14 16:52:33.788: INFO: Pod1 is running on kiedai4iakae-3. Tainting Node - Dec 14 16:52:34.004: INFO: Waiting up to 5m0s for pod "taint-eviction-b1" in namespace "taint-multiple-pods-5248" to be "running" - Dec 14 16:52:34.012: INFO: Pod "taint-eviction-b1": Phase="Pending", Reason="", readiness=false. Elapsed: 7.941817ms - Dec 14 16:52:36.019: INFO: Pod "taint-eviction-b1": Phase="Running", Reason="", readiness=true. Elapsed: 2.014613942s - Dec 14 16:52:36.019: INFO: Pod "taint-eviction-b1" satisfied condition "running" - Dec 14 16:52:36.019: INFO: Waiting up to 5m0s for pod "taint-eviction-b2" in namespace "taint-multiple-pods-5248" to be "running" - Dec 14 16:52:36.023: INFO: Pod "taint-eviction-b2": Phase="Running", Reason="", readiness=true. Elapsed: 4.660032ms - Dec 14 16:52:36.023: INFO: Pod "taint-eviction-b2" satisfied condition "running" - Dec 14 16:52:36.024: INFO: Pod2 is running on kiedai4iakae-3. Tainting Node - STEP: Trying to apply a taint on the Node 12/14/22 16:52:36.024 - STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 12/14/22 16:52:36.042 - STEP: Waiting for Pod1 and Pod2 to be deleted 12/14/22 16:52:36.048 - Dec 14 16:52:41.592: INFO: Noticed Pod "taint-eviction-b1" gets evicted. - Dec 14 16:53:01.632: INFO: Noticed Pod "taint-eviction-b2" gets evicted. - STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 12/14/22 16:53:01.65 - [AfterEach] [sig-node] NoExecuteTaintManager Multiple Pods [Serial] + STEP: Creating a kubernetes client 02/25/23 11:19:44.576 + Feb 25 11:19:44.576: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 11:19:44.581 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:19:44.617 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:19:44.632 + [It] should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:45 + STEP: Creating projection with secret that has name projected-secret-test-1efe4611-3488-4a7b-8f96-2296eb408b61 02/25/23 11:19:44.645 + STEP: Creating a pod to test consume secrets 02/25/23 11:19:44.668 + Feb 25 11:19:44.698: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302" in namespace "projected-8587" to be "Succeeded or Failed" + Feb 25 11:19:44.709: INFO: Pod "pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302": Phase="Pending", Reason="", readiness=false. Elapsed: 10.599493ms + Feb 25 11:19:46.717: INFO: Pod "pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018960503s + Feb 25 11:19:48.716: INFO: Pod "pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302": Phase="Pending", Reason="", readiness=false. Elapsed: 4.017035685s + Feb 25 11:19:50.717: INFO: Pod "pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.018719967s + STEP: Saw pod success 02/25/23 11:19:50.718 + Feb 25 11:19:50.718: INFO: Pod "pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302" satisfied condition "Succeeded or Failed" + Feb 25 11:19:50.723: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302 container projected-secret-volume-test: + STEP: delete the pod 02/25/23 11:19:50.736 + Feb 25 11:19:50.761: INFO: Waiting for pod pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302 to disappear + Feb 25 11:19:50.767: INFO: Pod pod-projected-secrets-55cedafd-8671-40d1-b1b7-6c058b725302 no longer exists + [AfterEach] [sig-storage] Projected secret test/e2e/framework/framework.go:187 - Dec 14 16:53:01.661: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "taint-multiple-pods-5248" for this suite. 12/14/22 16:53:01.668 - << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS + Feb 25 11:19:50.767: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-8587" for this suite. 02/25/23 11:19:50.773 + << End Captured GinkgoWriter Output ------------------------------ -[sig-node] Containers - should be able to override the image's default command and arguments [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:86 -[BeforeEach] [sig-node] Containers +[sig-storage] Subpath Atomic writer volumes + should support subpaths with downward pod [Conformance] + test/e2e/storage/subpath.go:92 +[BeforeEach] [sig-storage] Subpath test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:53:01.692 -Dec 14 16:53:01.692: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename containers 12/14/22 16:53:01.694 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:01.748 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:01.755 -[It] should be able to override the image's default command and arguments [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:86 -STEP: Creating a pod to test override all 12/14/22 16:53:01.759 -Dec 14 16:53:01.782: INFO: Waiting up to 5m0s for pod "client-containers-84df5696-36c1-4a65-a39c-167759887b10" in namespace "containers-906" to be "Succeeded or Failed" -Dec 14 16:53:01.791: INFO: Pod "client-containers-84df5696-36c1-4a65-a39c-167759887b10": Phase="Pending", Reason="", readiness=false. Elapsed: 9.285152ms -Dec 14 16:53:03.805: INFO: Pod "client-containers-84df5696-36c1-4a65-a39c-167759887b10": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023032465s -Dec 14 16:53:05.803: INFO: Pod "client-containers-84df5696-36c1-4a65-a39c-167759887b10": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021571629s -STEP: Saw pod success 12/14/22 16:53:05.803 -Dec 14 16:53:05.804: INFO: Pod "client-containers-84df5696-36c1-4a65-a39c-167759887b10" satisfied condition "Succeeded or Failed" -Dec 14 16:53:05.812: INFO: Trying to get logs from node kiedai4iakae-3 pod client-containers-84df5696-36c1-4a65-a39c-167759887b10 container agnhost-container: -STEP: delete the pod 12/14/22 16:53:06.206 -Dec 14 16:53:06.224: INFO: Waiting for pod client-containers-84df5696-36c1-4a65-a39c-167759887b10 to disappear -Dec 14 16:53:06.228: INFO: Pod client-containers-84df5696-36c1-4a65-a39c-167759887b10 no longer exists -[AfterEach] [sig-node] Containers +STEP: Creating a kubernetes client 02/25/23 11:19:50.785 +Feb 25 11:19:50.785: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename subpath 02/25/23 11:19:50.79 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:19:50.817 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:19:50.822 +[BeforeEach] Atomic writer volumes + test/e2e/storage/subpath.go:40 +STEP: Setting up data 02/25/23 11:19:50.826 +[It] should support subpaths with downward pod [Conformance] + test/e2e/storage/subpath.go:92 +STEP: Creating pod pod-subpath-test-downwardapi-4jzd 02/25/23 11:19:50.842 +STEP: Creating a pod to test atomic-volume-subpath 02/25/23 11:19:50.842 +Feb 25 11:19:50.857: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-4jzd" in namespace "subpath-2392" to be "Succeeded or Failed" +Feb 25 11:19:50.868: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Pending", Reason="", readiness=false. Elapsed: 11.007344ms +Feb 25 11:19:52.932: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.075092227s +Feb 25 11:19:54.876: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 4.018489461s +Feb 25 11:19:56.927: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 6.070079758s +Feb 25 11:19:58.888: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 8.031159459s +Feb 25 11:20:00.893: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 10.035352701s +Feb 25 11:20:02.875: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 12.017666588s +Feb 25 11:20:04.892: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 14.03511116s +Feb 25 11:20:06.893: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 16.035653271s +Feb 25 11:20:08.877: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 18.019834598s +Feb 25 11:20:10.903: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 20.045529676s +Feb 25 11:20:12.876: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 22.018900625s +Feb 25 11:20:14.895: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=false. Elapsed: 24.03824952s +Feb 25 11:20:16.877: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.019936886s +STEP: Saw pod success 02/25/23 11:20:16.877 +Feb 25 11:20:16.878: INFO: Pod "pod-subpath-test-downwardapi-4jzd" satisfied condition "Succeeded or Failed" +Feb 25 11:20:16.889: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-subpath-test-downwardapi-4jzd container test-container-subpath-downwardapi-4jzd: +STEP: delete the pod 02/25/23 11:20:16.9 +Feb 25 11:20:16.956: INFO: Waiting for pod pod-subpath-test-downwardapi-4jzd to disappear +Feb 25 11:20:16.962: INFO: Pod pod-subpath-test-downwardapi-4jzd no longer exists +STEP: Deleting pod pod-subpath-test-downwardapi-4jzd 02/25/23 11:20:16.962 +Feb 25 11:20:16.962: INFO: Deleting pod "pod-subpath-test-downwardapi-4jzd" in namespace "subpath-2392" +[AfterEach] [sig-storage] Subpath test/e2e/framework/framework.go:187 -Dec 14 16:53:06.228: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "containers-906" for this suite. 12/14/22 16:53:06.233 -{"msg":"PASSED [sig-node] Containers should be able to override the image's default command and arguments [NodeConformance] [Conformance]","completed":268,"skipped":4911,"failed":0} +Feb 25 11:20:16.969: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "subpath-2392" for this suite. 02/25/23 11:20:16.976 +{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with downward pod [Conformance]","completed":265,"skipped":4946,"failed":0} ------------------------------ -• [4.557 seconds] -[sig-node] Containers -test/e2e/common/node/framework.go:23 - should be able to override the image's default command and arguments [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:86 +• [SLOW TEST] [26.203 seconds] +[sig-storage] Subpath +test/e2e/storage/utils/framework.go:23 + Atomic writer volumes + test/e2e/storage/subpath.go:36 + should support subpaths with downward pod [Conformance] + test/e2e/storage/subpath.go:92 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Containers + [BeforeEach] [sig-storage] Subpath test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:53:01.692 - Dec 14 16:53:01.692: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename containers 12/14/22 16:53:01.694 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:01.748 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:01.755 - [It] should be able to override the image's default command and arguments [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:86 - STEP: Creating a pod to test override all 12/14/22 16:53:01.759 - Dec 14 16:53:01.782: INFO: Waiting up to 5m0s for pod "client-containers-84df5696-36c1-4a65-a39c-167759887b10" in namespace "containers-906" to be "Succeeded or Failed" - Dec 14 16:53:01.791: INFO: Pod "client-containers-84df5696-36c1-4a65-a39c-167759887b10": Phase="Pending", Reason="", readiness=false. Elapsed: 9.285152ms - Dec 14 16:53:03.805: INFO: Pod "client-containers-84df5696-36c1-4a65-a39c-167759887b10": Phase="Pending", Reason="", readiness=false. Elapsed: 2.023032465s - Dec 14 16:53:05.803: INFO: Pod "client-containers-84df5696-36c1-4a65-a39c-167759887b10": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.021571629s - STEP: Saw pod success 12/14/22 16:53:05.803 - Dec 14 16:53:05.804: INFO: Pod "client-containers-84df5696-36c1-4a65-a39c-167759887b10" satisfied condition "Succeeded or Failed" - Dec 14 16:53:05.812: INFO: Trying to get logs from node kiedai4iakae-3 pod client-containers-84df5696-36c1-4a65-a39c-167759887b10 container agnhost-container: - STEP: delete the pod 12/14/22 16:53:06.206 - Dec 14 16:53:06.224: INFO: Waiting for pod client-containers-84df5696-36c1-4a65-a39c-167759887b10 to disappear - Dec 14 16:53:06.228: INFO: Pod client-containers-84df5696-36c1-4a65-a39c-167759887b10 no longer exists - [AfterEach] [sig-node] Containers + STEP: Creating a kubernetes client 02/25/23 11:19:50.785 + Feb 25 11:19:50.785: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename subpath 02/25/23 11:19:50.79 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:19:50.817 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:19:50.822 + [BeforeEach] Atomic writer volumes + test/e2e/storage/subpath.go:40 + STEP: Setting up data 02/25/23 11:19:50.826 + [It] should support subpaths with downward pod [Conformance] + test/e2e/storage/subpath.go:92 + STEP: Creating pod pod-subpath-test-downwardapi-4jzd 02/25/23 11:19:50.842 + STEP: Creating a pod to test atomic-volume-subpath 02/25/23 11:19:50.842 + Feb 25 11:19:50.857: INFO: Waiting up to 5m0s for pod "pod-subpath-test-downwardapi-4jzd" in namespace "subpath-2392" to be "Succeeded or Failed" + Feb 25 11:19:50.868: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Pending", Reason="", readiness=false. Elapsed: 11.007344ms + Feb 25 11:19:52.932: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.075092227s + Feb 25 11:19:54.876: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 4.018489461s + Feb 25 11:19:56.927: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 6.070079758s + Feb 25 11:19:58.888: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 8.031159459s + Feb 25 11:20:00.893: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 10.035352701s + Feb 25 11:20:02.875: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 12.017666588s + Feb 25 11:20:04.892: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 14.03511116s + Feb 25 11:20:06.893: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 16.035653271s + Feb 25 11:20:08.877: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 18.019834598s + Feb 25 11:20:10.903: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 20.045529676s + Feb 25 11:20:12.876: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=true. Elapsed: 22.018900625s + Feb 25 11:20:14.895: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Running", Reason="", readiness=false. Elapsed: 24.03824952s + Feb 25 11:20:16.877: INFO: Pod "pod-subpath-test-downwardapi-4jzd": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.019936886s + STEP: Saw pod success 02/25/23 11:20:16.877 + Feb 25 11:20:16.878: INFO: Pod "pod-subpath-test-downwardapi-4jzd" satisfied condition "Succeeded or Failed" + Feb 25 11:20:16.889: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-subpath-test-downwardapi-4jzd container test-container-subpath-downwardapi-4jzd: + STEP: delete the pod 02/25/23 11:20:16.9 + Feb 25 11:20:16.956: INFO: Waiting for pod pod-subpath-test-downwardapi-4jzd to disappear + Feb 25 11:20:16.962: INFO: Pod pod-subpath-test-downwardapi-4jzd no longer exists + STEP: Deleting pod pod-subpath-test-downwardapi-4jzd 02/25/23 11:20:16.962 + Feb 25 11:20:16.962: INFO: Deleting pod "pod-subpath-test-downwardapi-4jzd" in namespace "subpath-2392" + [AfterEach] [sig-storage] Subpath test/e2e/framework/framework.go:187 - Dec 14 16:53:06.228: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "containers-906" for this suite. 12/14/22 16:53:06.233 + Feb 25 11:20:16.969: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "subpath-2392" for this suite. 02/25/23 11:20:16.976 << End Captured GinkgoWriter Output ------------------------------ -SSSS +SSS ------------------------------ -[sig-storage] Downward API volume - should provide container's cpu request [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:220 -[BeforeEach] [sig-storage] Downward API volume +[sig-api-machinery] Discovery + should validate PreferredVersion for each APIGroup [Conformance] + test/e2e/apimachinery/discovery.go:122 +[BeforeEach] [sig-api-machinery] Discovery test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:53:06.249 -Dec 14 16:53:06.249: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 16:53:06.252 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:06.276 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:06.279 -[BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 -[It] should provide container's cpu request [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:220 -STEP: Creating a pod to test downward API volume plugin 12/14/22 16:53:06.283 -Dec 14 16:53:06.301: INFO: Waiting up to 5m0s for pod "downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926" in namespace "downward-api-4587" to be "Succeeded or Failed" -Dec 14 16:53:06.305: INFO: Pod "downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926": Phase="Pending", Reason="", readiness=false. Elapsed: 4.187734ms -Dec 14 16:53:08.314: INFO: Pod "downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012905819s -Dec 14 16:53:10.316: INFO: Pod "downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014560217s -STEP: Saw pod success 12/14/22 16:53:10.316 -Dec 14 16:53:10.316: INFO: Pod "downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926" satisfied condition "Succeeded or Failed" -Dec 14 16:53:10.321: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926 container client-container: -STEP: delete the pod 12/14/22 16:53:10.332 -Dec 14 16:53:10.353: INFO: Waiting for pod downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926 to disappear -Dec 14 16:53:10.357: INFO: Pod downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926 no longer exists -[AfterEach] [sig-storage] Downward API volume +STEP: Creating a kubernetes client 02/25/23 11:20:16.995 +Feb 25 11:20:16.995: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename discovery 02/25/23 11:20:16.998 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:17.037 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:17.044 +[BeforeEach] [sig-api-machinery] Discovery + test/e2e/apimachinery/discovery.go:43 +STEP: Setting up server cert 02/25/23 11:20:17.05 +[It] should validate PreferredVersion for each APIGroup [Conformance] + test/e2e/apimachinery/discovery.go:122 +Feb 25 11:20:17.656: INFO: Checking APIGroup: apiregistration.k8s.io +Feb 25 11:20:17.657: INFO: PreferredVersion.GroupVersion: apiregistration.k8s.io/v1 +Feb 25 11:20:17.657: INFO: Versions found [{apiregistration.k8s.io/v1 v1}] +Feb 25 11:20:17.657: INFO: apiregistration.k8s.io/v1 matches apiregistration.k8s.io/v1 +Feb 25 11:20:17.657: INFO: Checking APIGroup: apps +Feb 25 11:20:17.659: INFO: PreferredVersion.GroupVersion: apps/v1 +Feb 25 11:20:17.659: INFO: Versions found [{apps/v1 v1}] +Feb 25 11:20:17.659: INFO: apps/v1 matches apps/v1 +Feb 25 11:20:17.659: INFO: Checking APIGroup: events.k8s.io +Feb 25 11:20:17.660: INFO: PreferredVersion.GroupVersion: events.k8s.io/v1 +Feb 25 11:20:17.660: INFO: Versions found [{events.k8s.io/v1 v1}] +Feb 25 11:20:17.660: INFO: events.k8s.io/v1 matches events.k8s.io/v1 +Feb 25 11:20:17.660: INFO: Checking APIGroup: authentication.k8s.io +Feb 25 11:20:17.661: INFO: PreferredVersion.GroupVersion: authentication.k8s.io/v1 +Feb 25 11:20:17.661: INFO: Versions found [{authentication.k8s.io/v1 v1}] +Feb 25 11:20:17.661: INFO: authentication.k8s.io/v1 matches authentication.k8s.io/v1 +Feb 25 11:20:17.661: INFO: Checking APIGroup: authorization.k8s.io +Feb 25 11:20:17.662: INFO: PreferredVersion.GroupVersion: authorization.k8s.io/v1 +Feb 25 11:20:17.662: INFO: Versions found [{authorization.k8s.io/v1 v1}] +Feb 25 11:20:17.662: INFO: authorization.k8s.io/v1 matches authorization.k8s.io/v1 +Feb 25 11:20:17.662: INFO: Checking APIGroup: autoscaling +Feb 25 11:20:17.663: INFO: PreferredVersion.GroupVersion: autoscaling/v2 +Feb 25 11:20:17.663: INFO: Versions found [{autoscaling/v2 v2} {autoscaling/v1 v1} {autoscaling/v2beta2 v2beta2}] +Feb 25 11:20:17.663: INFO: autoscaling/v2 matches autoscaling/v2 +Feb 25 11:20:17.663: INFO: Checking APIGroup: batch +Feb 25 11:20:17.665: INFO: PreferredVersion.GroupVersion: batch/v1 +Feb 25 11:20:17.665: INFO: Versions found [{batch/v1 v1}] +Feb 25 11:20:17.665: INFO: batch/v1 matches batch/v1 +Feb 25 11:20:17.665: INFO: Checking APIGroup: certificates.k8s.io +Feb 25 11:20:17.666: INFO: PreferredVersion.GroupVersion: certificates.k8s.io/v1 +Feb 25 11:20:17.666: INFO: Versions found [{certificates.k8s.io/v1 v1}] +Feb 25 11:20:17.666: INFO: certificates.k8s.io/v1 matches certificates.k8s.io/v1 +Feb 25 11:20:17.666: INFO: Checking APIGroup: networking.k8s.io +Feb 25 11:20:17.667: INFO: PreferredVersion.GroupVersion: networking.k8s.io/v1 +Feb 25 11:20:17.667: INFO: Versions found [{networking.k8s.io/v1 v1}] +Feb 25 11:20:17.667: INFO: networking.k8s.io/v1 matches networking.k8s.io/v1 +Feb 25 11:20:17.667: INFO: Checking APIGroup: policy +Feb 25 11:20:17.668: INFO: PreferredVersion.GroupVersion: policy/v1 +Feb 25 11:20:17.668: INFO: Versions found [{policy/v1 v1}] +Feb 25 11:20:17.668: INFO: policy/v1 matches policy/v1 +Feb 25 11:20:17.668: INFO: Checking APIGroup: rbac.authorization.k8s.io +Feb 25 11:20:17.670: INFO: PreferredVersion.GroupVersion: rbac.authorization.k8s.io/v1 +Feb 25 11:20:17.670: INFO: Versions found [{rbac.authorization.k8s.io/v1 v1}] +Feb 25 11:20:17.670: INFO: rbac.authorization.k8s.io/v1 matches rbac.authorization.k8s.io/v1 +Feb 25 11:20:17.670: INFO: Checking APIGroup: storage.k8s.io +Feb 25 11:20:17.673: INFO: PreferredVersion.GroupVersion: storage.k8s.io/v1 +Feb 25 11:20:17.673: INFO: Versions found [{storage.k8s.io/v1 v1} {storage.k8s.io/v1beta1 v1beta1}] +Feb 25 11:20:17.673: INFO: storage.k8s.io/v1 matches storage.k8s.io/v1 +Feb 25 11:20:17.673: INFO: Checking APIGroup: admissionregistration.k8s.io +Feb 25 11:20:17.675: INFO: PreferredVersion.GroupVersion: admissionregistration.k8s.io/v1 +Feb 25 11:20:17.675: INFO: Versions found [{admissionregistration.k8s.io/v1 v1}] +Feb 25 11:20:17.675: INFO: admissionregistration.k8s.io/v1 matches admissionregistration.k8s.io/v1 +Feb 25 11:20:17.675: INFO: Checking APIGroup: apiextensions.k8s.io +Feb 25 11:20:17.676: INFO: PreferredVersion.GroupVersion: apiextensions.k8s.io/v1 +Feb 25 11:20:17.676: INFO: Versions found [{apiextensions.k8s.io/v1 v1}] +Feb 25 11:20:17.677: INFO: apiextensions.k8s.io/v1 matches apiextensions.k8s.io/v1 +Feb 25 11:20:17.677: INFO: Checking APIGroup: scheduling.k8s.io +Feb 25 11:20:17.678: INFO: PreferredVersion.GroupVersion: scheduling.k8s.io/v1 +Feb 25 11:20:17.678: INFO: Versions found [{scheduling.k8s.io/v1 v1}] +Feb 25 11:20:17.678: INFO: scheduling.k8s.io/v1 matches scheduling.k8s.io/v1 +Feb 25 11:20:17.678: INFO: Checking APIGroup: coordination.k8s.io +Feb 25 11:20:17.680: INFO: PreferredVersion.GroupVersion: coordination.k8s.io/v1 +Feb 25 11:20:17.680: INFO: Versions found [{coordination.k8s.io/v1 v1}] +Feb 25 11:20:17.680: INFO: coordination.k8s.io/v1 matches coordination.k8s.io/v1 +Feb 25 11:20:17.680: INFO: Checking APIGroup: node.k8s.io +Feb 25 11:20:17.681: INFO: PreferredVersion.GroupVersion: node.k8s.io/v1 +Feb 25 11:20:17.681: INFO: Versions found [{node.k8s.io/v1 v1}] +Feb 25 11:20:17.681: INFO: node.k8s.io/v1 matches node.k8s.io/v1 +Feb 25 11:20:17.681: INFO: Checking APIGroup: discovery.k8s.io +Feb 25 11:20:17.682: INFO: PreferredVersion.GroupVersion: discovery.k8s.io/v1 +Feb 25 11:20:17.682: INFO: Versions found [{discovery.k8s.io/v1 v1}] +Feb 25 11:20:17.682: INFO: discovery.k8s.io/v1 matches discovery.k8s.io/v1 +Feb 25 11:20:17.682: INFO: Checking APIGroup: flowcontrol.apiserver.k8s.io +Feb 25 11:20:17.683: INFO: PreferredVersion.GroupVersion: flowcontrol.apiserver.k8s.io/v1beta2 +Feb 25 11:20:17.684: INFO: Versions found [{flowcontrol.apiserver.k8s.io/v1beta2 v1beta2} {flowcontrol.apiserver.k8s.io/v1beta1 v1beta1}] +Feb 25 11:20:17.684: INFO: flowcontrol.apiserver.k8s.io/v1beta2 matches flowcontrol.apiserver.k8s.io/v1beta2 +Feb 25 11:20:17.684: INFO: Checking APIGroup: cilium.io +Feb 25 11:20:17.685: INFO: PreferredVersion.GroupVersion: cilium.io/v2 +Feb 25 11:20:17.685: INFO: Versions found [{cilium.io/v2 v2} {cilium.io/v2alpha1 v2alpha1}] +Feb 25 11:20:17.685: INFO: cilium.io/v2 matches cilium.io/v2 +[AfterEach] [sig-api-machinery] Discovery test/e2e/framework/framework.go:187 -Dec 14 16:53:10.358: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-4587" for this suite. 12/14/22 16:53:10.363 -{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance]","completed":269,"skipped":4915,"failed":0} +Feb 25 11:20:17.685: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "discovery-9507" for this suite. 02/25/23 11:20:17.692 +{"msg":"PASSED [sig-api-machinery] Discovery should validate PreferredVersion for each APIGroup [Conformance]","completed":266,"skipped":4949,"failed":0} ------------------------------ -• [4.131 seconds] -[sig-storage] Downward API volume -test/e2e/common/storage/framework.go:23 - should provide container's cpu request [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:220 +• [0.712 seconds] +[sig-api-machinery] Discovery +test/e2e/apimachinery/framework.go:23 + should validate PreferredVersion for each APIGroup [Conformance] + test/e2e/apimachinery/discovery.go:122 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Downward API volume + [BeforeEach] [sig-api-machinery] Discovery test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:53:06.249 - Dec 14 16:53:06.249: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 16:53:06.252 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:06.276 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:06.279 - [BeforeEach] [sig-storage] Downward API volume - test/e2e/common/storage/downwardapi_volume.go:43 - [It] should provide container's cpu request [NodeConformance] [Conformance] - test/e2e/common/storage/downwardapi_volume.go:220 - STEP: Creating a pod to test downward API volume plugin 12/14/22 16:53:06.283 - Dec 14 16:53:06.301: INFO: Waiting up to 5m0s for pod "downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926" in namespace "downward-api-4587" to be "Succeeded or Failed" - Dec 14 16:53:06.305: INFO: Pod "downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926": Phase="Pending", Reason="", readiness=false. Elapsed: 4.187734ms - Dec 14 16:53:08.314: INFO: Pod "downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012905819s - Dec 14 16:53:10.316: INFO: Pod "downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.014560217s - STEP: Saw pod success 12/14/22 16:53:10.316 - Dec 14 16:53:10.316: INFO: Pod "downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926" satisfied condition "Succeeded or Failed" - Dec 14 16:53:10.321: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926 container client-container: - STEP: delete the pod 12/14/22 16:53:10.332 - Dec 14 16:53:10.353: INFO: Waiting for pod downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926 to disappear - Dec 14 16:53:10.357: INFO: Pod downwardapi-volume-5a2f9502-3821-43aa-86fb-106bc90df926 no longer exists - [AfterEach] [sig-storage] Downward API volume + STEP: Creating a kubernetes client 02/25/23 11:20:16.995 + Feb 25 11:20:16.995: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename discovery 02/25/23 11:20:16.998 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:17.037 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:17.044 + [BeforeEach] [sig-api-machinery] Discovery + test/e2e/apimachinery/discovery.go:43 + STEP: Setting up server cert 02/25/23 11:20:17.05 + [It] should validate PreferredVersion for each APIGroup [Conformance] + test/e2e/apimachinery/discovery.go:122 + Feb 25 11:20:17.656: INFO: Checking APIGroup: apiregistration.k8s.io + Feb 25 11:20:17.657: INFO: PreferredVersion.GroupVersion: apiregistration.k8s.io/v1 + Feb 25 11:20:17.657: INFO: Versions found [{apiregistration.k8s.io/v1 v1}] + Feb 25 11:20:17.657: INFO: apiregistration.k8s.io/v1 matches apiregistration.k8s.io/v1 + Feb 25 11:20:17.657: INFO: Checking APIGroup: apps + Feb 25 11:20:17.659: INFO: PreferredVersion.GroupVersion: apps/v1 + Feb 25 11:20:17.659: INFO: Versions found [{apps/v1 v1}] + Feb 25 11:20:17.659: INFO: apps/v1 matches apps/v1 + Feb 25 11:20:17.659: INFO: Checking APIGroup: events.k8s.io + Feb 25 11:20:17.660: INFO: PreferredVersion.GroupVersion: events.k8s.io/v1 + Feb 25 11:20:17.660: INFO: Versions found [{events.k8s.io/v1 v1}] + Feb 25 11:20:17.660: INFO: events.k8s.io/v1 matches events.k8s.io/v1 + Feb 25 11:20:17.660: INFO: Checking APIGroup: authentication.k8s.io + Feb 25 11:20:17.661: INFO: PreferredVersion.GroupVersion: authentication.k8s.io/v1 + Feb 25 11:20:17.661: INFO: Versions found [{authentication.k8s.io/v1 v1}] + Feb 25 11:20:17.661: INFO: authentication.k8s.io/v1 matches authentication.k8s.io/v1 + Feb 25 11:20:17.661: INFO: Checking APIGroup: authorization.k8s.io + Feb 25 11:20:17.662: INFO: PreferredVersion.GroupVersion: authorization.k8s.io/v1 + Feb 25 11:20:17.662: INFO: Versions found [{authorization.k8s.io/v1 v1}] + Feb 25 11:20:17.662: INFO: authorization.k8s.io/v1 matches authorization.k8s.io/v1 + Feb 25 11:20:17.662: INFO: Checking APIGroup: autoscaling + Feb 25 11:20:17.663: INFO: PreferredVersion.GroupVersion: autoscaling/v2 + Feb 25 11:20:17.663: INFO: Versions found [{autoscaling/v2 v2} {autoscaling/v1 v1} {autoscaling/v2beta2 v2beta2}] + Feb 25 11:20:17.663: INFO: autoscaling/v2 matches autoscaling/v2 + Feb 25 11:20:17.663: INFO: Checking APIGroup: batch + Feb 25 11:20:17.665: INFO: PreferredVersion.GroupVersion: batch/v1 + Feb 25 11:20:17.665: INFO: Versions found [{batch/v1 v1}] + Feb 25 11:20:17.665: INFO: batch/v1 matches batch/v1 + Feb 25 11:20:17.665: INFO: Checking APIGroup: certificates.k8s.io + Feb 25 11:20:17.666: INFO: PreferredVersion.GroupVersion: certificates.k8s.io/v1 + Feb 25 11:20:17.666: INFO: Versions found [{certificates.k8s.io/v1 v1}] + Feb 25 11:20:17.666: INFO: certificates.k8s.io/v1 matches certificates.k8s.io/v1 + Feb 25 11:20:17.666: INFO: Checking APIGroup: networking.k8s.io + Feb 25 11:20:17.667: INFO: PreferredVersion.GroupVersion: networking.k8s.io/v1 + Feb 25 11:20:17.667: INFO: Versions found [{networking.k8s.io/v1 v1}] + Feb 25 11:20:17.667: INFO: networking.k8s.io/v1 matches networking.k8s.io/v1 + Feb 25 11:20:17.667: INFO: Checking APIGroup: policy + Feb 25 11:20:17.668: INFO: PreferredVersion.GroupVersion: policy/v1 + Feb 25 11:20:17.668: INFO: Versions found [{policy/v1 v1}] + Feb 25 11:20:17.668: INFO: policy/v1 matches policy/v1 + Feb 25 11:20:17.668: INFO: Checking APIGroup: rbac.authorization.k8s.io + Feb 25 11:20:17.670: INFO: PreferredVersion.GroupVersion: rbac.authorization.k8s.io/v1 + Feb 25 11:20:17.670: INFO: Versions found [{rbac.authorization.k8s.io/v1 v1}] + Feb 25 11:20:17.670: INFO: rbac.authorization.k8s.io/v1 matches rbac.authorization.k8s.io/v1 + Feb 25 11:20:17.670: INFO: Checking APIGroup: storage.k8s.io + Feb 25 11:20:17.673: INFO: PreferredVersion.GroupVersion: storage.k8s.io/v1 + Feb 25 11:20:17.673: INFO: Versions found [{storage.k8s.io/v1 v1} {storage.k8s.io/v1beta1 v1beta1}] + Feb 25 11:20:17.673: INFO: storage.k8s.io/v1 matches storage.k8s.io/v1 + Feb 25 11:20:17.673: INFO: Checking APIGroup: admissionregistration.k8s.io + Feb 25 11:20:17.675: INFO: PreferredVersion.GroupVersion: admissionregistration.k8s.io/v1 + Feb 25 11:20:17.675: INFO: Versions found [{admissionregistration.k8s.io/v1 v1}] + Feb 25 11:20:17.675: INFO: admissionregistration.k8s.io/v1 matches admissionregistration.k8s.io/v1 + Feb 25 11:20:17.675: INFO: Checking APIGroup: apiextensions.k8s.io + Feb 25 11:20:17.676: INFO: PreferredVersion.GroupVersion: apiextensions.k8s.io/v1 + Feb 25 11:20:17.676: INFO: Versions found [{apiextensions.k8s.io/v1 v1}] + Feb 25 11:20:17.677: INFO: apiextensions.k8s.io/v1 matches apiextensions.k8s.io/v1 + Feb 25 11:20:17.677: INFO: Checking APIGroup: scheduling.k8s.io + Feb 25 11:20:17.678: INFO: PreferredVersion.GroupVersion: scheduling.k8s.io/v1 + Feb 25 11:20:17.678: INFO: Versions found [{scheduling.k8s.io/v1 v1}] + Feb 25 11:20:17.678: INFO: scheduling.k8s.io/v1 matches scheduling.k8s.io/v1 + Feb 25 11:20:17.678: INFO: Checking APIGroup: coordination.k8s.io + Feb 25 11:20:17.680: INFO: PreferredVersion.GroupVersion: coordination.k8s.io/v1 + Feb 25 11:20:17.680: INFO: Versions found [{coordination.k8s.io/v1 v1}] + Feb 25 11:20:17.680: INFO: coordination.k8s.io/v1 matches coordination.k8s.io/v1 + Feb 25 11:20:17.680: INFO: Checking APIGroup: node.k8s.io + Feb 25 11:20:17.681: INFO: PreferredVersion.GroupVersion: node.k8s.io/v1 + Feb 25 11:20:17.681: INFO: Versions found [{node.k8s.io/v1 v1}] + Feb 25 11:20:17.681: INFO: node.k8s.io/v1 matches node.k8s.io/v1 + Feb 25 11:20:17.681: INFO: Checking APIGroup: discovery.k8s.io + Feb 25 11:20:17.682: INFO: PreferredVersion.GroupVersion: discovery.k8s.io/v1 + Feb 25 11:20:17.682: INFO: Versions found [{discovery.k8s.io/v1 v1}] + Feb 25 11:20:17.682: INFO: discovery.k8s.io/v1 matches discovery.k8s.io/v1 + Feb 25 11:20:17.682: INFO: Checking APIGroup: flowcontrol.apiserver.k8s.io + Feb 25 11:20:17.683: INFO: PreferredVersion.GroupVersion: flowcontrol.apiserver.k8s.io/v1beta2 + Feb 25 11:20:17.684: INFO: Versions found [{flowcontrol.apiserver.k8s.io/v1beta2 v1beta2} {flowcontrol.apiserver.k8s.io/v1beta1 v1beta1}] + Feb 25 11:20:17.684: INFO: flowcontrol.apiserver.k8s.io/v1beta2 matches flowcontrol.apiserver.k8s.io/v1beta2 + Feb 25 11:20:17.684: INFO: Checking APIGroup: cilium.io + Feb 25 11:20:17.685: INFO: PreferredVersion.GroupVersion: cilium.io/v2 + Feb 25 11:20:17.685: INFO: Versions found [{cilium.io/v2 v2} {cilium.io/v2alpha1 v2alpha1}] + Feb 25 11:20:17.685: INFO: cilium.io/v2 matches cilium.io/v2 + [AfterEach] [sig-api-machinery] Discovery test/e2e/framework/framework.go:187 - Dec 14 16:53:10.358: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-4587" for this suite. 12/14/22 16:53:10.363 + Feb 25 11:20:17.685: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "discovery-9507" for this suite. 02/25/23 11:20:17.692 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SS ------------------------------ -[sig-network] DNS - should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] - test/e2e/network/dns.go:193 -[BeforeEach] [sig-network] DNS +[sig-api-machinery] Garbage collector + should delete RS created by deployment when not orphaning [Conformance] + test/e2e/apimachinery/garbage_collector.go:491 +[BeforeEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:53:10.389 -Dec 14 16:53:10.389: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename dns 12/14/22 16:53:10.391 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:10.428 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:10.434 -[It] should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] - test/e2e/network/dns.go:193 -STEP: Creating a test headless service 12/14/22 16:53:10.438 -STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-9692 A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-9692;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-9692 A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-9692;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-9692.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-9692.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-9692.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-9692.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-9692.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-9692.svc;check="$$(dig +notcp +noall +answer +search 56.40.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.40.56_udp@PTR;check="$$(dig +tcp +noall +answer +search 56.40.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.40.56_tcp@PTR;sleep 1; done - 12/14/22 16:53:10.479 -STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-9692 A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-9692;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-9692 A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-9692;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-9692.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-9692.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-9692.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-9692.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-9692.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-9692.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-9692.svc;check="$$(dig +notcp +noall +answer +search 56.40.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.40.56_udp@PTR;check="$$(dig +tcp +noall +answer +search 56.40.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.40.56_tcp@PTR;sleep 1; done - 12/14/22 16:53:10.479 -STEP: creating a pod to probe DNS 12/14/22 16:53:10.48 -STEP: submitting the pod to kubernetes 12/14/22 16:53:10.48 -Dec 14 16:53:10.500: INFO: Waiting up to 15m0s for pod "dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a" in namespace "dns-9692" to be "running" -Dec 14 16:53:10.505: INFO: Pod "dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.527096ms -Dec 14 16:53:12.513: INFO: Pod "dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012828069s -Dec 14 16:53:14.513: INFO: Pod "dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a": Phase="Running", Reason="", readiness=true. Elapsed: 4.012508535s -Dec 14 16:53:14.513: INFO: Pod "dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a" satisfied condition "running" -STEP: retrieving the pod 12/14/22 16:53:14.513 -STEP: looking for the results for each expected name from probers 12/14/22 16:53:14.521 -Dec 14 16:53:14.528: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.533: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.538: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.542: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.546: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.552: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.559: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.564: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.593: INFO: Unable to read jessie_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.598: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.604: INFO: Unable to read jessie_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.609: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.614: INFO: Unable to read jessie_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.619: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.626: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.631: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:14.652: INFO: Lookups using dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-9692 wheezy_tcp@dns-test-service.dns-9692 wheezy_udp@dns-test-service.dns-9692.svc wheezy_tcp@dns-test-service.dns-9692.svc wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-9692 jessie_tcp@dns-test-service.dns-9692 jessie_udp@dns-test-service.dns-9692.svc jessie_tcp@dns-test-service.dns-9692.svc jessie_udp@_http._tcp.dns-test-service.dns-9692.svc jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc] - -Dec 14 16:53:19.660: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.664: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.674: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.681: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.685: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.689: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.693: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.697: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.722: INFO: Unable to read jessie_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.725: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.729: INFO: Unable to read jessie_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.733: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.737: INFO: Unable to read jessie_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.741: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.745: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.749: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:19.775: INFO: Lookups using dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-9692 wheezy_tcp@dns-test-service.dns-9692 wheezy_udp@dns-test-service.dns-9692.svc wheezy_tcp@dns-test-service.dns-9692.svc wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-9692 jessie_tcp@dns-test-service.dns-9692 jessie_udp@dns-test-service.dns-9692.svc jessie_tcp@dns-test-service.dns-9692.svc jessie_udp@_http._tcp.dns-test-service.dns-9692.svc jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc] - -Dec 14 16:53:24.663: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.669: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.677: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.685: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.691: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.701: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.710: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.717: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.758: INFO: Unable to read jessie_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.773: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.785: INFO: Unable to read jessie_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.791: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.795: INFO: Unable to read jessie_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.800: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.812: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.820: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:24.874: INFO: Lookups using dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-9692 wheezy_tcp@dns-test-service.dns-9692 wheezy_udp@dns-test-service.dns-9692.svc wheezy_tcp@dns-test-service.dns-9692.svc wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-9692 jessie_tcp@dns-test-service.dns-9692 jessie_udp@dns-test-service.dns-9692.svc jessie_tcp@dns-test-service.dns-9692.svc jessie_udp@_http._tcp.dns-test-service.dns-9692.svc jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc] - -Dec 14 16:53:29.664: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.670: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.677: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.684: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.689: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.695: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.699: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.704: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.725: INFO: Unable to read jessie_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.729: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.734: INFO: Unable to read jessie_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.739: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.743: INFO: Unable to read jessie_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.748: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.755: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.782: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:29.809: INFO: Lookups using dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-9692 wheezy_tcp@dns-test-service.dns-9692 wheezy_udp@dns-test-service.dns-9692.svc wheezy_tcp@dns-test-service.dns-9692.svc wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-9692 jessie_tcp@dns-test-service.dns-9692 jessie_udp@dns-test-service.dns-9692.svc jessie_tcp@dns-test-service.dns-9692.svc jessie_udp@_http._tcp.dns-test-service.dns-9692.svc jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc] - -Dec 14 16:53:34.663: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.670: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.676: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.682: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.690: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.696: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.709: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.715: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.743: INFO: Unable to read jessie_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.747: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.760: INFO: Unable to read jessie_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.766: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.777: INFO: Unable to read jessie_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.784: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.792: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.797: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:34.824: INFO: Lookups using dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-9692 wheezy_tcp@dns-test-service.dns-9692 wheezy_udp@dns-test-service.dns-9692.svc wheezy_tcp@dns-test-service.dns-9692.svc wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-9692 jessie_tcp@dns-test-service.dns-9692 jessie_udp@dns-test-service.dns-9692.svc jessie_tcp@dns-test-service.dns-9692.svc jessie_udp@_http._tcp.dns-test-service.dns-9692.svc jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc] - -Dec 14 16:53:39.660: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.665: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.673: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.678: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.685: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.690: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.694: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.701: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.726: INFO: Unable to read jessie_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.729: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.734: INFO: Unable to read jessie_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.739: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.742: INFO: Unable to read jessie_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.748: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.768: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.783: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) -Dec 14 16:53:39.814: INFO: Lookups using dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-9692 wheezy_tcp@dns-test-service.dns-9692 wheezy_udp@dns-test-service.dns-9692.svc wheezy_tcp@dns-test-service.dns-9692.svc wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-9692 jessie_tcp@dns-test-service.dns-9692 jessie_udp@dns-test-service.dns-9692.svc jessie_tcp@dns-test-service.dns-9692.svc jessie_udp@_http._tcp.dns-test-service.dns-9692.svc jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc] - -Dec 14 16:53:44.770: INFO: DNS probes using dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a succeeded - -STEP: deleting the pod 12/14/22 16:53:44.77 -STEP: deleting the test service 12/14/22 16:53:44.799 -STEP: deleting the test headless service 12/14/22 16:53:44.886 -[AfterEach] [sig-network] DNS +STEP: Creating a kubernetes client 02/25/23 11:20:17.708 +Feb 25 11:20:17.709: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename gc 02/25/23 11:20:17.712 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:17.736 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:17.74 +[It] should delete RS created by deployment when not orphaning [Conformance] + test/e2e/apimachinery/garbage_collector.go:491 +STEP: create the deployment 02/25/23 11:20:17.744 +STEP: Wait for the Deployment to create new ReplicaSet 02/25/23 11:20:17.756 +STEP: delete the deployment 02/25/23 11:20:18.269 +STEP: wait for all rs to be garbage collected 02/25/23 11:20:18.297 +STEP: expected 0 rs, got 1 rs 02/25/23 11:20:18.306 +STEP: expected 0 pods, got 2 pods 02/25/23 11:20:18.318 +STEP: Gathering metrics 02/25/23 11:20:18.852 +Feb 25 11:20:19.156: INFO: Waiting up to 5m0s for pod "kube-controller-manager-diek3dua9gay-2" in namespace "kube-system" to be "running and ready" +Feb 25 11:20:19.166: INFO: Pod "kube-controller-manager-diek3dua9gay-2": Phase="Running", Reason="", readiness=true. Elapsed: 10.000109ms +Feb 25 11:20:19.166: INFO: The phase of Pod kube-controller-manager-diek3dua9gay-2 is Running (Ready = true) +Feb 25 11:20:19.167: INFO: Pod "kube-controller-manager-diek3dua9gay-2" satisfied condition "running and ready" +Feb 25 11:20:19.290: INFO: For apiserver_request_total: +For apiserver_request_latency_seconds: +For apiserver_init_events_total: +For garbage_collector_attempt_to_delete_queue_latency: +For garbage_collector_attempt_to_delete_work_duration: +For garbage_collector_attempt_to_orphan_queue_latency: +For garbage_collector_attempt_to_orphan_work_duration: +For garbage_collector_dirty_processing_latency_microseconds: +For garbage_collector_event_processing_latency_microseconds: +For garbage_collector_graph_changes_queue_latency: +For garbage_collector_graph_changes_work_duration: +For garbage_collector_orphan_processing_latency_microseconds: +For namespace_queue_latency: +For namespace_queue_latency_sum: +For namespace_queue_latency_count: +For namespace_retries: +For namespace_work_duration: +For namespace_work_duration_sum: +For namespace_work_duration_count: +For function_duration_seconds: +For errors_total: +For evicted_pods_total: + +[AfterEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:187 -Dec 14 16:53:44.911: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "dns-9692" for this suite. 12/14/22 16:53:44.925 -{"msg":"PASSED [sig-network] DNS should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance]","completed":270,"skipped":4948,"failed":0} +Feb 25 11:20:19.291: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "gc-396" for this suite. 02/25/23 11:20:19.315 +{"msg":"PASSED [sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance]","completed":267,"skipped":4951,"failed":0} ------------------------------ -• [SLOW TEST] [34.561 seconds] -[sig-network] DNS -test/e2e/network/common/framework.go:23 - should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] - test/e2e/network/dns.go:193 +• [1.618 seconds] +[sig-api-machinery] Garbage collector +test/e2e/apimachinery/framework.go:23 + should delete RS created by deployment when not orphaning [Conformance] + test/e2e/apimachinery/garbage_collector.go:491 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] DNS + [BeforeEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:53:10.389 - Dec 14 16:53:10.389: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename dns 12/14/22 16:53:10.391 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:10.428 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:10.434 - [It] should resolve DNS of partial qualified names for services [LinuxOnly] [Conformance] - test/e2e/network/dns.go:193 - STEP: Creating a test headless service 12/14/22 16:53:10.438 - STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-9692 A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-9692;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-9692 A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-9692;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-9692.svc A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-9692.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-9692.svc A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-9692.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-9692.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-9692.svc;check="$$(dig +notcp +noall +answer +search 56.40.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.40.56_udp@PTR;check="$$(dig +tcp +noall +answer +search 56.40.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.40.56_tcp@PTR;sleep 1; done - 12/14/22 16:53:10.479 - STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service;check="$$(dig +tcp +noall +answer +search dns-test-service A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-9692 A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-9692;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-9692 A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-9692;check="$$(dig +notcp +noall +answer +search dns-test-service.dns-9692.svc A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-9692.svc;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-9692.svc A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-9692.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-9692.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-9692.svc;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-9692.svc SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-9692.svc;check="$$(dig +notcp +noall +answer +search 56.40.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.40.56_udp@PTR;check="$$(dig +tcp +noall +answer +search 56.40.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.40.56_tcp@PTR;sleep 1; done - 12/14/22 16:53:10.479 - STEP: creating a pod to probe DNS 12/14/22 16:53:10.48 - STEP: submitting the pod to kubernetes 12/14/22 16:53:10.48 - Dec 14 16:53:10.500: INFO: Waiting up to 15m0s for pod "dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a" in namespace "dns-9692" to be "running" - Dec 14 16:53:10.505: INFO: Pod "dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.527096ms - Dec 14 16:53:12.513: INFO: Pod "dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012828069s - Dec 14 16:53:14.513: INFO: Pod "dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a": Phase="Running", Reason="", readiness=true. Elapsed: 4.012508535s - Dec 14 16:53:14.513: INFO: Pod "dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a" satisfied condition "running" - STEP: retrieving the pod 12/14/22 16:53:14.513 - STEP: looking for the results for each expected name from probers 12/14/22 16:53:14.521 - Dec 14 16:53:14.528: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.533: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.538: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.542: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.546: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.552: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.559: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.564: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.593: INFO: Unable to read jessie_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.598: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.604: INFO: Unable to read jessie_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.609: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.614: INFO: Unable to read jessie_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.619: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.626: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.631: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:14.652: INFO: Lookups using dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-9692 wheezy_tcp@dns-test-service.dns-9692 wheezy_udp@dns-test-service.dns-9692.svc wheezy_tcp@dns-test-service.dns-9692.svc wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-9692 jessie_tcp@dns-test-service.dns-9692 jessie_udp@dns-test-service.dns-9692.svc jessie_tcp@dns-test-service.dns-9692.svc jessie_udp@_http._tcp.dns-test-service.dns-9692.svc jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc] - - Dec 14 16:53:19.660: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.664: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.674: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.681: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.685: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.689: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.693: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.697: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.722: INFO: Unable to read jessie_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.725: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.729: INFO: Unable to read jessie_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.733: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.737: INFO: Unable to read jessie_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.741: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.745: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.749: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:19.775: INFO: Lookups using dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-9692 wheezy_tcp@dns-test-service.dns-9692 wheezy_udp@dns-test-service.dns-9692.svc wheezy_tcp@dns-test-service.dns-9692.svc wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-9692 jessie_tcp@dns-test-service.dns-9692 jessie_udp@dns-test-service.dns-9692.svc jessie_tcp@dns-test-service.dns-9692.svc jessie_udp@_http._tcp.dns-test-service.dns-9692.svc jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc] - - Dec 14 16:53:24.663: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.669: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.677: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.685: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.691: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.701: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.710: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.717: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.758: INFO: Unable to read jessie_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.773: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.785: INFO: Unable to read jessie_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.791: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.795: INFO: Unable to read jessie_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.800: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.812: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.820: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:24.874: INFO: Lookups using dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-9692 wheezy_tcp@dns-test-service.dns-9692 wheezy_udp@dns-test-service.dns-9692.svc wheezy_tcp@dns-test-service.dns-9692.svc wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-9692 jessie_tcp@dns-test-service.dns-9692 jessie_udp@dns-test-service.dns-9692.svc jessie_tcp@dns-test-service.dns-9692.svc jessie_udp@_http._tcp.dns-test-service.dns-9692.svc jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc] - - Dec 14 16:53:29.664: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.670: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.677: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.684: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.689: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.695: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.699: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.704: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.725: INFO: Unable to read jessie_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.729: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.734: INFO: Unable to read jessie_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.739: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.743: INFO: Unable to read jessie_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.748: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.755: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.782: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:29.809: INFO: Lookups using dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-9692 wheezy_tcp@dns-test-service.dns-9692 wheezy_udp@dns-test-service.dns-9692.svc wheezy_tcp@dns-test-service.dns-9692.svc wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-9692 jessie_tcp@dns-test-service.dns-9692 jessie_udp@dns-test-service.dns-9692.svc jessie_tcp@dns-test-service.dns-9692.svc jessie_udp@_http._tcp.dns-test-service.dns-9692.svc jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc] - - Dec 14 16:53:34.663: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.670: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.676: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.682: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.690: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.696: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.709: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.715: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.743: INFO: Unable to read jessie_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.747: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.760: INFO: Unable to read jessie_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.766: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.777: INFO: Unable to read jessie_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.784: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.792: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.797: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:34.824: INFO: Lookups using dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-9692 wheezy_tcp@dns-test-service.dns-9692 wheezy_udp@dns-test-service.dns-9692.svc wheezy_tcp@dns-test-service.dns-9692.svc wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-9692 jessie_tcp@dns-test-service.dns-9692 jessie_udp@dns-test-service.dns-9692.svc jessie_tcp@dns-test-service.dns-9692.svc jessie_udp@_http._tcp.dns-test-service.dns-9692.svc jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc] - - Dec 14 16:53:39.660: INFO: Unable to read wheezy_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.665: INFO: Unable to read wheezy_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.673: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.678: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.685: INFO: Unable to read wheezy_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.690: INFO: Unable to read wheezy_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.694: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.701: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.726: INFO: Unable to read jessie_udp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.729: INFO: Unable to read jessie_tcp@dns-test-service from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.734: INFO: Unable to read jessie_udp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.739: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692 from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.742: INFO: Unable to read jessie_udp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.748: INFO: Unable to read jessie_tcp@dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.768: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.783: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc from pod dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a: the server could not find the requested resource (get pods dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a) - Dec 14 16:53:39.814: INFO: Lookups using dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a failed for: [wheezy_udp@dns-test-service wheezy_tcp@dns-test-service wheezy_udp@dns-test-service.dns-9692 wheezy_tcp@dns-test-service.dns-9692 wheezy_udp@dns-test-service.dns-9692.svc wheezy_tcp@dns-test-service.dns-9692.svc wheezy_udp@_http._tcp.dns-test-service.dns-9692.svc wheezy_tcp@_http._tcp.dns-test-service.dns-9692.svc jessie_udp@dns-test-service jessie_tcp@dns-test-service jessie_udp@dns-test-service.dns-9692 jessie_tcp@dns-test-service.dns-9692 jessie_udp@dns-test-service.dns-9692.svc jessie_tcp@dns-test-service.dns-9692.svc jessie_udp@_http._tcp.dns-test-service.dns-9692.svc jessie_tcp@_http._tcp.dns-test-service.dns-9692.svc] - - Dec 14 16:53:44.770: INFO: DNS probes using dns-9692/dns-test-c87c1df6-8fdf-4adf-81bc-8a1c4cd06c2a succeeded - - STEP: deleting the pod 12/14/22 16:53:44.77 - STEP: deleting the test service 12/14/22 16:53:44.799 - STEP: deleting the test headless service 12/14/22 16:53:44.886 - [AfterEach] [sig-network] DNS + STEP: Creating a kubernetes client 02/25/23 11:20:17.708 + Feb 25 11:20:17.709: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename gc 02/25/23 11:20:17.712 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:17.736 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:17.74 + [It] should delete RS created by deployment when not orphaning [Conformance] + test/e2e/apimachinery/garbage_collector.go:491 + STEP: create the deployment 02/25/23 11:20:17.744 + STEP: Wait for the Deployment to create new ReplicaSet 02/25/23 11:20:17.756 + STEP: delete the deployment 02/25/23 11:20:18.269 + STEP: wait for all rs to be garbage collected 02/25/23 11:20:18.297 + STEP: expected 0 rs, got 1 rs 02/25/23 11:20:18.306 + STEP: expected 0 pods, got 2 pods 02/25/23 11:20:18.318 + STEP: Gathering metrics 02/25/23 11:20:18.852 + Feb 25 11:20:19.156: INFO: Waiting up to 5m0s for pod "kube-controller-manager-diek3dua9gay-2" in namespace "kube-system" to be "running and ready" + Feb 25 11:20:19.166: INFO: Pod "kube-controller-manager-diek3dua9gay-2": Phase="Running", Reason="", readiness=true. Elapsed: 10.000109ms + Feb 25 11:20:19.166: INFO: The phase of Pod kube-controller-manager-diek3dua9gay-2 is Running (Ready = true) + Feb 25 11:20:19.167: INFO: Pod "kube-controller-manager-diek3dua9gay-2" satisfied condition "running and ready" + Feb 25 11:20:19.290: INFO: For apiserver_request_total: + For apiserver_request_latency_seconds: + For apiserver_init_events_total: + For garbage_collector_attempt_to_delete_queue_latency: + For garbage_collector_attempt_to_delete_work_duration: + For garbage_collector_attempt_to_orphan_queue_latency: + For garbage_collector_attempt_to_orphan_work_duration: + For garbage_collector_dirty_processing_latency_microseconds: + For garbage_collector_event_processing_latency_microseconds: + For garbage_collector_graph_changes_queue_latency: + For garbage_collector_graph_changes_work_duration: + For garbage_collector_orphan_processing_latency_microseconds: + For namespace_queue_latency: + For namespace_queue_latency_sum: + For namespace_queue_latency_count: + For namespace_retries: + For namespace_work_duration: + For namespace_work_duration_sum: + For namespace_work_duration_count: + For function_duration_seconds: + For errors_total: + For evicted_pods_total: + + [AfterEach] [sig-api-machinery] Garbage collector test/e2e/framework/framework.go:187 - Dec 14 16:53:44.911: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "dns-9692" for this suite. 12/14/22 16:53:44.925 + Feb 25 11:20:19.291: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "gc-396" for this suite. 02/25/23 11:20:19.315 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSS ------------------------------ -[sig-cli] Kubectl client Proxy server - should support --unix-socket=/path [Conformance] - test/e2e/kubectl/kubectl.go:1810 -[BeforeEach] [sig-cli] Kubectl client +[sig-node] PreStop + should call prestop when killing a pod [Conformance] + test/e2e/node/pre_stop.go:168 +[BeforeEach] [sig-node] PreStop test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:53:44.955 -Dec 14 16:53:44.955: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 16:53:44.959 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:44.991 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:44.996 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 -[It] should support --unix-socket=/path [Conformance] - test/e2e/kubectl/kubectl.go:1810 -STEP: Starting the proxy 12/14/22 16:53:45.001 -Dec 14 16:53:45.009: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2340 proxy --unix-socket=/tmp/kubectl-proxy-unix1368532952/test' -STEP: retrieving proxy /api/ output 12/14/22 16:53:45.204 -[AfterEach] [sig-cli] Kubectl client +STEP: Creating a kubernetes client 02/25/23 11:20:19.333 +Feb 25 11:20:19.333: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename prestop 02/25/23 11:20:19.341 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:19.384 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:19.389 +[BeforeEach] [sig-node] PreStop + test/e2e/node/pre_stop.go:159 +[It] should call prestop when killing a pod [Conformance] + test/e2e/node/pre_stop.go:168 +STEP: Creating server pod server in namespace prestop-4964 02/25/23 11:20:19.403 +STEP: Waiting for pods to come up. 02/25/23 11:20:19.419 +Feb 25 11:20:19.420: INFO: Waiting up to 5m0s for pod "server" in namespace "prestop-4964" to be "running" +Feb 25 11:20:19.459: INFO: Pod "server": Phase="Pending", Reason="", readiness=false. Elapsed: 39.230289ms +Feb 25 11:20:21.466: INFO: Pod "server": Phase="Running", Reason="", readiness=true. Elapsed: 2.046182094s +Feb 25 11:20:21.466: INFO: Pod "server" satisfied condition "running" +STEP: Creating tester pod tester in namespace prestop-4964 02/25/23 11:20:21.472 +Feb 25 11:20:21.482: INFO: Waiting up to 5m0s for pod "tester" in namespace "prestop-4964" to be "running" +Feb 25 11:20:21.490: INFO: Pod "tester": Phase="Pending", Reason="", readiness=false. Elapsed: 8.384367ms +Feb 25 11:20:23.506: INFO: Pod "tester": Phase="Running", Reason="", readiness=true. Elapsed: 2.024575772s +Feb 25 11:20:23.506: INFO: Pod "tester" satisfied condition "running" +STEP: Deleting pre-stop pod 02/25/23 11:20:23.506 +Feb 25 11:20:28.534: INFO: Saw: { + "Hostname": "server", + "Sent": null, + "Received": { + "prestop": 1 + }, + "Errors": null, + "Log": [ + "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", + "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up." + ], + "StillContactingPeers": true +} +STEP: Deleting the server pod 02/25/23 11:20:28.534 +[AfterEach] [sig-node] PreStop test/e2e/framework/framework.go:187 -Dec 14 16:53:45.208: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-2340" for this suite. 12/14/22 16:53:45.216 -{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support --unix-socket=/path [Conformance]","completed":271,"skipped":4977,"failed":0} +Feb 25 11:20:28.572: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "prestop-4964" for this suite. 02/25/23 11:20:28.59 +{"msg":"PASSED [sig-node] PreStop should call prestop when killing a pod [Conformance]","completed":268,"skipped":4959,"failed":0} ------------------------------ -• [0.276 seconds] -[sig-cli] Kubectl client -test/e2e/kubectl/framework.go:23 - Proxy server - test/e2e/kubectl/kubectl.go:1778 - should support --unix-socket=/path [Conformance] - test/e2e/kubectl/kubectl.go:1810 +• [SLOW TEST] [9.272 seconds] +[sig-node] PreStop +test/e2e/node/framework.go:23 + should call prestop when killing a pod [Conformance] + test/e2e/node/pre_stop.go:168 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-cli] Kubectl client + [BeforeEach] [sig-node] PreStop test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:53:44.955 - Dec 14 16:53:44.955: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 16:53:44.959 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:44.991 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:44.996 - [BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 - [It] should support --unix-socket=/path [Conformance] - test/e2e/kubectl/kubectl.go:1810 - STEP: Starting the proxy 12/14/22 16:53:45.001 - Dec 14 16:53:45.009: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2340 proxy --unix-socket=/tmp/kubectl-proxy-unix1368532952/test' - STEP: retrieving proxy /api/ output 12/14/22 16:53:45.204 - [AfterEach] [sig-cli] Kubectl client + STEP: Creating a kubernetes client 02/25/23 11:20:19.333 + Feb 25 11:20:19.333: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename prestop 02/25/23 11:20:19.341 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:19.384 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:19.389 + [BeforeEach] [sig-node] PreStop + test/e2e/node/pre_stop.go:159 + [It] should call prestop when killing a pod [Conformance] + test/e2e/node/pre_stop.go:168 + STEP: Creating server pod server in namespace prestop-4964 02/25/23 11:20:19.403 + STEP: Waiting for pods to come up. 02/25/23 11:20:19.419 + Feb 25 11:20:19.420: INFO: Waiting up to 5m0s for pod "server" in namespace "prestop-4964" to be "running" + Feb 25 11:20:19.459: INFO: Pod "server": Phase="Pending", Reason="", readiness=false. Elapsed: 39.230289ms + Feb 25 11:20:21.466: INFO: Pod "server": Phase="Running", Reason="", readiness=true. Elapsed: 2.046182094s + Feb 25 11:20:21.466: INFO: Pod "server" satisfied condition "running" + STEP: Creating tester pod tester in namespace prestop-4964 02/25/23 11:20:21.472 + Feb 25 11:20:21.482: INFO: Waiting up to 5m0s for pod "tester" in namespace "prestop-4964" to be "running" + Feb 25 11:20:21.490: INFO: Pod "tester": Phase="Pending", Reason="", readiness=false. Elapsed: 8.384367ms + Feb 25 11:20:23.506: INFO: Pod "tester": Phase="Running", Reason="", readiness=true. Elapsed: 2.024575772s + Feb 25 11:20:23.506: INFO: Pod "tester" satisfied condition "running" + STEP: Deleting pre-stop pod 02/25/23 11:20:23.506 + Feb 25 11:20:28.534: INFO: Saw: { + "Hostname": "server", + "Sent": null, + "Received": { + "prestop": 1 + }, + "Errors": null, + "Log": [ + "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up.", + "default/nettest has 0 endpoints ([]), which is less than 8 as expected. Waiting for all endpoints to come up." + ], + "StillContactingPeers": true + } + STEP: Deleting the server pod 02/25/23 11:20:28.534 + [AfterEach] [sig-node] PreStop test/e2e/framework/framework.go:187 - Dec 14 16:53:45.208: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-2340" for this suite. 12/14/22 16:53:45.216 + Feb 25 11:20:28.572: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "prestop-4964" for this suite. 02/25/23 11:20:28.59 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-apps] ReplicationController - should adopt matching pods on creation [Conformance] - test/e2e/apps/rc.go:91 -[BeforeEach] [sig-apps] ReplicationController +[sig-node] Pods + should delete a collection of pods [Conformance] + test/e2e/common/node/pods.go:844 +[BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:53:45.235 -Dec 14 16:53:45.235: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename replication-controller 12/14/22 16:53:45.236 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:45.31 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:45.314 -[BeforeEach] [sig-apps] ReplicationController - test/e2e/apps/rc.go:56 -[It] should adopt matching pods on creation [Conformance] - test/e2e/apps/rc.go:91 -STEP: Given a Pod with a 'name' label pod-adoption is created 12/14/22 16:53:45.323 -Dec 14 16:53:45.339: INFO: Waiting up to 5m0s for pod "pod-adoption" in namespace "replication-controller-2073" to be "running and ready" -Dec 14 16:53:45.347: INFO: Pod "pod-adoption": Phase="Pending", Reason="", readiness=false. Elapsed: 8.456351ms -Dec 14 16:53:45.348: INFO: The phase of Pod pod-adoption is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:53:47.355: INFO: Pod "pod-adoption": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015951168s -Dec 14 16:53:47.355: INFO: The phase of Pod pod-adoption is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:53:49.359: INFO: Pod "pod-adoption": Phase="Running", Reason="", readiness=true. Elapsed: 4.019906565s -Dec 14 16:53:49.359: INFO: The phase of Pod pod-adoption is Running (Ready = true) -Dec 14 16:53:49.359: INFO: Pod "pod-adoption" satisfied condition "running and ready" -STEP: When a replication controller with a matching selector is created 12/14/22 16:53:49.366 -STEP: Then the orphan pod is adopted 12/14/22 16:53:49.375 -[AfterEach] [sig-apps] ReplicationController +STEP: Creating a kubernetes client 02/25/23 11:20:28.614 +Feb 25 11:20:28.614: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename pods 02/25/23 11:20:28.619 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:28.644 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:28.65 +[BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 +[It] should delete a collection of pods [Conformance] + test/e2e/common/node/pods.go:844 +STEP: Create set of pods 02/25/23 11:20:28.654 +Feb 25 11:20:28.667: INFO: created test-pod-1 +Feb 25 11:20:28.696: INFO: created test-pod-2 +Feb 25 11:20:28.710: INFO: created test-pod-3 +STEP: waiting for all 3 pods to be running 02/25/23 11:20:28.71 +Feb 25 11:20:28.710: INFO: Waiting up to 5m0s for all pods (need at least 3) in namespace 'pods-6386' to be running and ready +Feb 25 11:20:28.732: INFO: The status of Pod test-pod-1 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed +Feb 25 11:20:28.732: INFO: The status of Pod test-pod-2 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed +Feb 25 11:20:28.732: INFO: The status of Pod test-pod-3 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed +Feb 25 11:20:28.732: INFO: 0 / 3 pods in namespace 'pods-6386' are running and ready (0 seconds elapsed) +Feb 25 11:20:28.732: INFO: expected 0 pod replicas in namespace 'pods-6386', 0 are Running and Ready. +Feb 25 11:20:28.732: INFO: POD NODE PHASE GRACE CONDITIONS +Feb 25 11:20:28.732: INFO: test-pod-1 diek3dua9gay-3 Pending [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:20:28 +0000 UTC }] +Feb 25 11:20:28.732: INFO: test-pod-2 diek3dua9gay-2 Pending [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:20:28 +0000 UTC }] +Feb 25 11:20:28.732: INFO: test-pod-3 diek3dua9gay-3 Pending [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:20:28 +0000 UTC }] +Feb 25 11:20:28.732: INFO: +Feb 25 11:20:30.749: INFO: 3 / 3 pods in namespace 'pods-6386' are running and ready (2 seconds elapsed) +Feb 25 11:20:30.749: INFO: expected 0 pod replicas in namespace 'pods-6386', 0 are Running and Ready. +STEP: waiting for all pods to be deleted 02/25/23 11:20:30.786 +Feb 25 11:20:30.792: INFO: Pod quantity 3 is different from expected quantity 0 +Feb 25 11:20:31.800: INFO: Pod quantity 3 is different from expected quantity 0 +Feb 25 11:20:32.800: INFO: Pod quantity 3 is different from expected quantity 0 +Feb 25 11:20:33.801: INFO: Pod quantity 1 is different from expected quantity 0 +[AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 -Dec 14 16:53:50.385: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "replication-controller-2073" for this suite. 12/14/22 16:53:50.403 -{"msg":"PASSED [sig-apps] ReplicationController should adopt matching pods on creation [Conformance]","completed":272,"skipped":5015,"failed":0} +Feb 25 11:20:34.802: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-6386" for this suite. 02/25/23 11:20:34.81 +{"msg":"PASSED [sig-node] Pods should delete a collection of pods [Conformance]","completed":269,"skipped":4981,"failed":0} ------------------------------ -• [SLOW TEST] [5.183 seconds] -[sig-apps] ReplicationController -test/e2e/apps/framework.go:23 - should adopt matching pods on creation [Conformance] - test/e2e/apps/rc.go:91 +• [SLOW TEST] [6.208 seconds] +[sig-node] Pods +test/e2e/common/node/framework.go:23 + should delete a collection of pods [Conformance] + test/e2e/common/node/pods.go:844 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] ReplicationController - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:53:45.235 - Dec 14 16:53:45.235: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename replication-controller 12/14/22 16:53:45.236 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:45.31 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:45.314 - [BeforeEach] [sig-apps] ReplicationController - test/e2e/apps/rc.go:56 - [It] should adopt matching pods on creation [Conformance] - test/e2e/apps/rc.go:91 - STEP: Given a Pod with a 'name' label pod-adoption is created 12/14/22 16:53:45.323 - Dec 14 16:53:45.339: INFO: Waiting up to 5m0s for pod "pod-adoption" in namespace "replication-controller-2073" to be "running and ready" - Dec 14 16:53:45.347: INFO: Pod "pod-adoption": Phase="Pending", Reason="", readiness=false. Elapsed: 8.456351ms - Dec 14 16:53:45.348: INFO: The phase of Pod pod-adoption is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:53:47.355: INFO: Pod "pod-adoption": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015951168s - Dec 14 16:53:47.355: INFO: The phase of Pod pod-adoption is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:53:49.359: INFO: Pod "pod-adoption": Phase="Running", Reason="", readiness=true. Elapsed: 4.019906565s - Dec 14 16:53:49.359: INFO: The phase of Pod pod-adoption is Running (Ready = true) - Dec 14 16:53:49.359: INFO: Pod "pod-adoption" satisfied condition "running and ready" - STEP: When a replication controller with a matching selector is created 12/14/22 16:53:49.366 - STEP: Then the orphan pod is adopted 12/14/22 16:53:49.375 - [AfterEach] [sig-apps] ReplicationController + [BeforeEach] [sig-node] Pods + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 11:20:28.614 + Feb 25 11:20:28.614: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename pods 02/25/23 11:20:28.619 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:28.644 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:28.65 + [BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 + [It] should delete a collection of pods [Conformance] + test/e2e/common/node/pods.go:844 + STEP: Create set of pods 02/25/23 11:20:28.654 + Feb 25 11:20:28.667: INFO: created test-pod-1 + Feb 25 11:20:28.696: INFO: created test-pod-2 + Feb 25 11:20:28.710: INFO: created test-pod-3 + STEP: waiting for all 3 pods to be running 02/25/23 11:20:28.71 + Feb 25 11:20:28.710: INFO: Waiting up to 5m0s for all pods (need at least 3) in namespace 'pods-6386' to be running and ready + Feb 25 11:20:28.732: INFO: The status of Pod test-pod-1 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed + Feb 25 11:20:28.732: INFO: The status of Pod test-pod-2 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed + Feb 25 11:20:28.732: INFO: The status of Pod test-pod-3 is Pending (Ready = false), waiting for it to be either Running (with Ready = true) or Failed + Feb 25 11:20:28.732: INFO: 0 / 3 pods in namespace 'pods-6386' are running and ready (0 seconds elapsed) + Feb 25 11:20:28.732: INFO: expected 0 pod replicas in namespace 'pods-6386', 0 are Running and Ready. + Feb 25 11:20:28.732: INFO: POD NODE PHASE GRACE CONDITIONS + Feb 25 11:20:28.732: INFO: test-pod-1 diek3dua9gay-3 Pending [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:20:28 +0000 UTC }] + Feb 25 11:20:28.732: INFO: test-pod-2 diek3dua9gay-2 Pending [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:20:28 +0000 UTC }] + Feb 25 11:20:28.732: INFO: test-pod-3 diek3dua9gay-3 Pending [{PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:20:28 +0000 UTC }] + Feb 25 11:20:28.732: INFO: + Feb 25 11:20:30.749: INFO: 3 / 3 pods in namespace 'pods-6386' are running and ready (2 seconds elapsed) + Feb 25 11:20:30.749: INFO: expected 0 pod replicas in namespace 'pods-6386', 0 are Running and Ready. + STEP: waiting for all pods to be deleted 02/25/23 11:20:30.786 + Feb 25 11:20:30.792: INFO: Pod quantity 3 is different from expected quantity 0 + Feb 25 11:20:31.800: INFO: Pod quantity 3 is different from expected quantity 0 + Feb 25 11:20:32.800: INFO: Pod quantity 3 is different from expected quantity 0 + Feb 25 11:20:33.801: INFO: Pod quantity 1 is different from expected quantity 0 + [AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 - Dec 14 16:53:50.385: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "replication-controller-2073" for this suite. 12/14/22 16:53:50.403 + Feb 25 11:20:34.802: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "pods-6386" for this suite. 02/25/23 11:20:34.81 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-network] Ingress API - should support creating Ingress API operations [Conformance] - test/e2e/network/ingress.go:552 -[BeforeEach] [sig-network] Ingress API +[sig-cli] Kubectl client Proxy server + should support proxy with --port 0 [Conformance] + test/e2e/kubectl/kubectl.go:1785 +[BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:53:50.423 -Dec 14 16:53:50.423: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename ingress 12/14/22 16:53:50.426 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:50.448 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:50.452 -[It] should support creating Ingress API operations [Conformance] - test/e2e/network/ingress.go:552 -STEP: getting /apis 12/14/22 16:53:50.458 -STEP: getting /apis/networking.k8s.io 12/14/22 16:53:50.462 -STEP: getting /apis/networking.k8s.iov1 12/14/22 16:53:50.463 -STEP: creating 12/14/22 16:53:50.465 -STEP: getting 12/14/22 16:53:50.51 -STEP: listing 12/14/22 16:53:50.514 -STEP: watching 12/14/22 16:53:50.518 -Dec 14 16:53:50.518: INFO: starting watch -STEP: cluster-wide listing 12/14/22 16:53:50.519 -STEP: cluster-wide watching 12/14/22 16:53:50.533 -Dec 14 16:53:50.533: INFO: starting watch -STEP: patching 12/14/22 16:53:50.535 -STEP: updating 12/14/22 16:53:50.55 -Dec 14 16:53:50.565: INFO: waiting for watch events with expected annotations -Dec 14 16:53:50.565: INFO: saw patched and updated annotations -STEP: patching /status 12/14/22 16:53:50.565 -STEP: updating /status 12/14/22 16:53:50.573 -STEP: get /status 12/14/22 16:53:50.584 -STEP: deleting 12/14/22 16:53:50.592 -STEP: deleting a collection 12/14/22 16:53:50.614 -[AfterEach] [sig-network] Ingress API +STEP: Creating a kubernetes client 02/25/23 11:20:34.852 +Feb 25 11:20:34.853: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 11:20:34.856 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:34.899 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:34.92 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 +[It] should support proxy with --port 0 [Conformance] + test/e2e/kubectl/kubectl.go:1785 +STEP: starting the proxy server 02/25/23 11:20:34.924 +Feb 25 11:20:34.927: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5869 proxy -p 0 --disable-filter' +STEP: curling proxy /api/ output 02/25/23 11:20:35.141 +[AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 -Dec 14 16:53:50.645: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "ingress-3427" for this suite. 12/14/22 16:53:50.653 -{"msg":"PASSED [sig-network] Ingress API should support creating Ingress API operations [Conformance]","completed":273,"skipped":5036,"failed":0} +Feb 25 11:20:35.163: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-5869" for this suite. 02/25/23 11:20:35.17 +{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support proxy with --port 0 [Conformance]","completed":270,"skipped":5106,"failed":0} ------------------------------ -• [0.241 seconds] -[sig-network] Ingress API -test/e2e/network/common/framework.go:23 - should support creating Ingress API operations [Conformance] - test/e2e/network/ingress.go:552 +• [0.326 seconds] +[sig-cli] Kubectl client +test/e2e/kubectl/framework.go:23 + Proxy server + test/e2e/kubectl/kubectl.go:1778 + should support proxy with --port 0 [Conformance] + test/e2e/kubectl/kubectl.go:1785 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Ingress API + [BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:53:50.423 - Dec 14 16:53:50.423: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename ingress 12/14/22 16:53:50.426 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:50.448 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:50.452 - [It] should support creating Ingress API operations [Conformance] - test/e2e/network/ingress.go:552 - STEP: getting /apis 12/14/22 16:53:50.458 - STEP: getting /apis/networking.k8s.io 12/14/22 16:53:50.462 - STEP: getting /apis/networking.k8s.iov1 12/14/22 16:53:50.463 - STEP: creating 12/14/22 16:53:50.465 - STEP: getting 12/14/22 16:53:50.51 - STEP: listing 12/14/22 16:53:50.514 - STEP: watching 12/14/22 16:53:50.518 - Dec 14 16:53:50.518: INFO: starting watch - STEP: cluster-wide listing 12/14/22 16:53:50.519 - STEP: cluster-wide watching 12/14/22 16:53:50.533 - Dec 14 16:53:50.533: INFO: starting watch - STEP: patching 12/14/22 16:53:50.535 - STEP: updating 12/14/22 16:53:50.55 - Dec 14 16:53:50.565: INFO: waiting for watch events with expected annotations - Dec 14 16:53:50.565: INFO: saw patched and updated annotations - STEP: patching /status 12/14/22 16:53:50.565 - STEP: updating /status 12/14/22 16:53:50.573 - STEP: get /status 12/14/22 16:53:50.584 - STEP: deleting 12/14/22 16:53:50.592 - STEP: deleting a collection 12/14/22 16:53:50.614 - [AfterEach] [sig-network] Ingress API + STEP: Creating a kubernetes client 02/25/23 11:20:34.852 + Feb 25 11:20:34.853: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 11:20:34.856 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:34.899 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:34.92 + [BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 + [It] should support proxy with --port 0 [Conformance] + test/e2e/kubectl/kubectl.go:1785 + STEP: starting the proxy server 02/25/23 11:20:34.924 + Feb 25 11:20:34.927: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5869 proxy -p 0 --disable-filter' + STEP: curling proxy /api/ output 02/25/23 11:20:35.141 + [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 16:53:50.645: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "ingress-3427" for this suite. 12/14/22 16:53:50.653 + Feb 25 11:20:35.163: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-5869" for this suite. 02/25/23 11:20:35.17 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSS +SSSSSSSSSSSSSSS ------------------------------ [sig-auth] ServiceAccounts - should guarantee kube-root-ca.crt exist in any namespace [Conformance] - test/e2e/auth/service_accounts.go:739 + should mount projected service account token [Conformance] + test/e2e/auth/service_accounts.go:272 [BeforeEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:53:50.668 -Dec 14 16:53:50.668: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename svcaccounts 12/14/22 16:53:50.67 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:50.727 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:50.737 -[It] should guarantee kube-root-ca.crt exist in any namespace [Conformance] - test/e2e/auth/service_accounts.go:739 -Dec 14 16:53:50.760: INFO: Got root ca configmap in namespace "svcaccounts-6805" -Dec 14 16:53:50.772: INFO: Deleted root ca configmap in namespace "svcaccounts-6805" -STEP: waiting for a new root ca configmap created 12/14/22 16:53:51.273 -Dec 14 16:53:51.279: INFO: Recreated root ca configmap in namespace "svcaccounts-6805" -Dec 14 16:53:51.288: INFO: Updated root ca configmap in namespace "svcaccounts-6805" -STEP: waiting for the root ca configmap reconciled 12/14/22 16:53:51.788 -Dec 14 16:53:51.795: INFO: Reconciled root ca configmap in namespace "svcaccounts-6805" +STEP: Creating a kubernetes client 02/25/23 11:20:35.179 +Feb 25 11:20:35.179: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename svcaccounts 02/25/23 11:20:35.181 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:35.215 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:35.222 +[It] should mount projected service account token [Conformance] + test/e2e/auth/service_accounts.go:272 +STEP: Creating a pod to test service account token: 02/25/23 11:20:35.227 +Feb 25 11:20:35.243: INFO: Waiting up to 5m0s for pod "test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0" in namespace "svcaccounts-9092" to be "Succeeded or Failed" +Feb 25 11:20:35.248: INFO: Pod "test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0": Phase="Pending", Reason="", readiness=false. Elapsed: 5.174042ms +Feb 25 11:20:37.256: INFO: Pod "test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013172675s +Feb 25 11:20:39.254: INFO: Pod "test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0": Phase="Pending", Reason="", readiness=false. Elapsed: 4.010898676s +Feb 25 11:20:41.258: INFO: Pod "test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.015456247s +STEP: Saw pod success 02/25/23 11:20:41.258 +Feb 25 11:20:41.259: INFO: Pod "test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0" satisfied condition "Succeeded or Failed" +Feb 25 11:20:41.264: INFO: Trying to get logs from node diek3dua9gay-3 pod test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0 container agnhost-container: +STEP: delete the pod 02/25/23 11:20:41.274 +Feb 25 11:20:41.305: INFO: Waiting for pod test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0 to disappear +Feb 25 11:20:41.311: INFO: Pod test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0 no longer exists [AfterEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:187 -Dec 14 16:53:51.796: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "svcaccounts-6805" for this suite. 12/14/22 16:53:51.802 -{"msg":"PASSED [sig-auth] ServiceAccounts should guarantee kube-root-ca.crt exist in any namespace [Conformance]","completed":274,"skipped":5050,"failed":0} +Feb 25 11:20:41.312: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "svcaccounts-9092" for this suite. 02/25/23 11:20:41.32 +{"msg":"PASSED [sig-auth] ServiceAccounts should mount projected service account token [Conformance]","completed":271,"skipped":5121,"failed":0} ------------------------------ -• [1.143 seconds] +• [SLOW TEST] [6.156 seconds] [sig-auth] ServiceAccounts test/e2e/auth/framework.go:23 - should guarantee kube-root-ca.crt exist in any namespace [Conformance] - test/e2e/auth/service_accounts.go:739 + should mount projected service account token [Conformance] + test/e2e/auth/service_accounts.go:272 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:53:50.668 - Dec 14 16:53:50.668: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename svcaccounts 12/14/22 16:53:50.67 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:50.727 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:50.737 - [It] should guarantee kube-root-ca.crt exist in any namespace [Conformance] - test/e2e/auth/service_accounts.go:739 - Dec 14 16:53:50.760: INFO: Got root ca configmap in namespace "svcaccounts-6805" - Dec 14 16:53:50.772: INFO: Deleted root ca configmap in namespace "svcaccounts-6805" - STEP: waiting for a new root ca configmap created 12/14/22 16:53:51.273 - Dec 14 16:53:51.279: INFO: Recreated root ca configmap in namespace "svcaccounts-6805" - Dec 14 16:53:51.288: INFO: Updated root ca configmap in namespace "svcaccounts-6805" - STEP: waiting for the root ca configmap reconciled 12/14/22 16:53:51.788 - Dec 14 16:53:51.795: INFO: Reconciled root ca configmap in namespace "svcaccounts-6805" + STEP: Creating a kubernetes client 02/25/23 11:20:35.179 + Feb 25 11:20:35.179: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename svcaccounts 02/25/23 11:20:35.181 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:35.215 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:35.222 + [It] should mount projected service account token [Conformance] + test/e2e/auth/service_accounts.go:272 + STEP: Creating a pod to test service account token: 02/25/23 11:20:35.227 + Feb 25 11:20:35.243: INFO: Waiting up to 5m0s for pod "test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0" in namespace "svcaccounts-9092" to be "Succeeded or Failed" + Feb 25 11:20:35.248: INFO: Pod "test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0": Phase="Pending", Reason="", readiness=false. Elapsed: 5.174042ms + Feb 25 11:20:37.256: INFO: Pod "test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013172675s + Feb 25 11:20:39.254: INFO: Pod "test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0": Phase="Pending", Reason="", readiness=false. Elapsed: 4.010898676s + Feb 25 11:20:41.258: INFO: Pod "test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.015456247s + STEP: Saw pod success 02/25/23 11:20:41.258 + Feb 25 11:20:41.259: INFO: Pod "test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0" satisfied condition "Succeeded or Failed" + Feb 25 11:20:41.264: INFO: Trying to get logs from node diek3dua9gay-3 pod test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0 container agnhost-container: + STEP: delete the pod 02/25/23 11:20:41.274 + Feb 25 11:20:41.305: INFO: Waiting for pod test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0 to disappear + Feb 25 11:20:41.311: INFO: Pod test-pod-4ebdef94-7ab0-49aa-a140-d74933fc23b0 no longer exists [AfterEach] [sig-auth] ServiceAccounts test/e2e/framework/framework.go:187 - Dec 14 16:53:51.796: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "svcaccounts-6805" for this suite. 12/14/22 16:53:51.802 + Feb 25 11:20:41.312: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "svcaccounts-9092" for this suite. 02/25/23 11:20:41.32 << End Captured GinkgoWriter Output ------------------------------ -[sig-storage] ConfigMap - should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:88 -[BeforeEach] [sig-storage] ConfigMap +SSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-network] Service endpoints latency + should not be very high [Conformance] + test/e2e/network/service_latency.go:59 +[BeforeEach] [sig-network] Service endpoints latency test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:53:51.813 -Dec 14 16:53:51.813: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename configmap 12/14/22 16:53:51.816 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:51.841 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:51.847 -[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:88 -STEP: Creating configMap with name configmap-test-volume-map-d156e7c6-a92d-4769-9bff-2e1c1a797d87 12/14/22 16:53:51.867 -STEP: Creating a pod to test consume configMaps 12/14/22 16:53:51.873 -Dec 14 16:53:51.885: INFO: Waiting up to 5m0s for pod "pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786" in namespace "configmap-8437" to be "Succeeded or Failed" -Dec 14 16:53:51.908: INFO: Pod "pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786": Phase="Pending", Reason="", readiness=false. Elapsed: 22.545311ms -Dec 14 16:53:53.919: INFO: Pod "pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786": Phase="Running", Reason="", readiness=true. Elapsed: 2.03396165s -Dec 14 16:53:55.915: INFO: Pod "pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786": Phase="Running", Reason="", readiness=false. Elapsed: 4.029530009s -Dec 14 16:53:57.915: INFO: Pod "pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.029901115s -STEP: Saw pod success 12/14/22 16:53:57.916 -Dec 14 16:53:57.916: INFO: Pod "pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786" satisfied condition "Succeeded or Failed" -Dec 14 16:53:57.923: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786 container agnhost-container: -STEP: delete the pod 12/14/22 16:53:58.217 -Dec 14 16:53:58.234: INFO: Waiting for pod pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786 to disappear -Dec 14 16:53:58.238: INFO: Pod pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786 no longer exists -[AfterEach] [sig-storage] ConfigMap +STEP: Creating a kubernetes client 02/25/23 11:20:41.341 +Feb 25 11:20:41.341: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename svc-latency 02/25/23 11:20:41.342 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:41.371 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:41.377 +[It] should not be very high [Conformance] + test/e2e/network/service_latency.go:59 +Feb 25 11:20:41.382: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: creating replication controller svc-latency-rc in namespace svc-latency-4048 02/25/23 11:20:41.384 +I0225 11:20:41.409426 13 runners.go:193] Created replication controller with name: svc-latency-rc, namespace: svc-latency-4048, replica count: 1 +I0225 11:20:42.483420 13 runners.go:193] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0225 11:20:43.483760 13 runners.go:193] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +I0225 11:20:44.484295 13 runners.go:193] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Feb 25 11:20:44.608: INFO: Created: latency-svc-gd5mr +Feb 25 11:20:44.673: INFO: Got endpoints: latency-svc-gd5mr [88.42354ms] +Feb 25 11:20:44.723: INFO: Created: latency-svc-vbjtm +Feb 25 11:20:44.729: INFO: Created: latency-svc-dthdf +Feb 25 11:20:44.735: INFO: Got endpoints: latency-svc-dthdf [60.695301ms] +Feb 25 11:20:44.762: INFO: Created: latency-svc-dwkqx +Feb 25 11:20:44.766: INFO: Got endpoints: latency-svc-vbjtm [91.0181ms] +Feb 25 11:20:44.789: INFO: Created: latency-svc-xxwmb +Feb 25 11:20:44.808: INFO: Got endpoints: latency-svc-dwkqx [132.993021ms] +Feb 25 11:20:44.820: INFO: Created: latency-svc-v2lwj +Feb 25 11:20:44.823: INFO: Got endpoints: latency-svc-xxwmb [147.591886ms] +Feb 25 11:20:44.832: INFO: Got endpoints: latency-svc-v2lwj [156.861217ms] +Feb 25 11:20:44.837: INFO: Created: latency-svc-7xtlf +Feb 25 11:20:44.865: INFO: Got endpoints: latency-svc-7xtlf [188.803702ms] +Feb 25 11:20:44.922: INFO: Created: latency-svc-4t8lk +Feb 25 11:20:44.956: INFO: Created: latency-svc-gbxlz +Feb 25 11:20:44.977: INFO: Got endpoints: latency-svc-4t8lk [300.913142ms] +Feb 25 11:20:44.990: INFO: Created: latency-svc-7mpmp +Feb 25 11:20:45.006: INFO: Got endpoints: latency-svc-gbxlz [330.08416ms] +Feb 25 11:20:45.055: INFO: Got endpoints: latency-svc-7mpmp [378.509026ms] +Feb 25 11:20:45.085: INFO: Created: latency-svc-gtqj5 +Feb 25 11:20:45.133: INFO: Got endpoints: latency-svc-gtqj5 [456.128882ms] +Feb 25 11:20:45.153: INFO: Created: latency-svc-vmbmh +Feb 25 11:20:45.176: INFO: Created: latency-svc-tjxct +Feb 25 11:20:45.176: INFO: Got endpoints: latency-svc-tjxct [499.948498ms] +Feb 25 11:20:45.194: INFO: Got endpoints: latency-svc-vmbmh [517.548535ms] +Feb 25 11:20:45.438: INFO: Created: latency-svc-9662q +Feb 25 11:20:45.438: INFO: Created: latency-svc-dmgxp +Feb 25 11:20:45.439: INFO: Created: latency-svc-pbpbq +Feb 25 11:20:45.439: INFO: Created: latency-svc-9c2nf +Feb 25 11:20:45.439: INFO: Created: latency-svc-dkhfn +Feb 25 11:20:45.440: INFO: Created: latency-svc-kjvsl +Feb 25 11:20:45.440: INFO: Created: latency-svc-2r4kx +Feb 25 11:20:45.454: INFO: Created: latency-svc-gt48c +Feb 25 11:20:45.454: INFO: Created: latency-svc-csr4k +Feb 25 11:20:45.454: INFO: Created: latency-svc-kqvgc +Feb 25 11:20:45.516: INFO: Created: latency-svc-fxpbs +Feb 25 11:20:45.517: INFO: Created: latency-svc-wjm5p +Feb 25 11:20:45.517: INFO: Created: latency-svc-5v2s9 +Feb 25 11:20:45.530: INFO: Got endpoints: latency-svc-9c2nf [353.01572ms] +Feb 25 11:20:45.530: INFO: Got endpoints: latency-svc-kjvsl [706.883466ms] +Feb 25 11:20:45.531: INFO: Got endpoints: latency-svc-dkhfn [397.483689ms] +Feb 25 11:20:45.531: INFO: Got endpoints: latency-svc-dmgxp [854.284573ms] +Feb 25 11:20:45.531: INFO: Got endpoints: latency-svc-pbpbq [336.955898ms] +Feb 25 11:20:45.554: INFO: Created: latency-svc-xdvss +Feb 25 11:20:45.555: INFO: Created: latency-svc-nqcc4 +Feb 25 11:20:45.581: INFO: Got endpoints: latency-svc-2r4kx [902.843147ms] +Feb 25 11:20:45.582: INFO: Got endpoints: latency-svc-gt48c [905.540065ms] +Feb 25 11:20:45.619: INFO: Got endpoints: latency-svc-csr4k [810.286136ms] +Feb 25 11:20:45.651: INFO: Got endpoints: latency-svc-9662q [915.840616ms] +Feb 25 11:20:45.661: INFO: Got endpoints: latency-svc-kqvgc [654.40904ms] +Feb 25 11:20:45.662: INFO: Got endpoints: latency-svc-wjm5p [796.903663ms] +Feb 25 11:20:45.695: INFO: Got endpoints: latency-svc-5v2s9 [928.629602ms] +Feb 25 11:20:45.695: INFO: Got endpoints: latency-svc-fxpbs [862.561566ms] +Feb 25 11:20:45.699: INFO: Created: latency-svc-872bc +Feb 25 11:20:45.726: INFO: Got endpoints: latency-svc-nqcc4 [670.400875ms] +Feb 25 11:20:45.733: INFO: Got endpoints: latency-svc-872bc [202.787455ms] +Feb 25 11:20:45.733: INFO: Got endpoints: latency-svc-xdvss [755.854578ms] +Feb 25 11:20:45.907: INFO: Created: latency-svc-wctmd +Feb 25 11:20:45.921: INFO: Created: latency-svc-8hr6z +Feb 25 11:20:45.923: INFO: Created: latency-svc-rwk65 +Feb 25 11:20:45.926: INFO: Created: latency-svc-nnq2h +Feb 25 11:20:45.926: INFO: Created: latency-svc-mjl62 +Feb 25 11:20:45.957: INFO: Got endpoints: latency-svc-wctmd [223.755658ms] +Feb 25 11:20:45.968: INFO: Created: latency-svc-f5gql +Feb 25 11:20:45.968: INFO: Created: latency-svc-47hfb +Feb 25 11:20:45.969: INFO: Created: latency-svc-69jhh +Feb 25 11:20:45.969: INFO: Created: latency-svc-fm7sm +Feb 25 11:20:45.969: INFO: Created: latency-svc-2j8p4 +Feb 25 11:20:45.974: INFO: Created: latency-svc-57bff +Feb 25 11:20:45.974: INFO: Created: latency-svc-dpcg4 +Feb 25 11:20:45.974: INFO: Created: latency-svc-ztmw7 +Feb 25 11:20:45.975: INFO: Created: latency-svc-tjnxl +Feb 25 11:20:45.975: INFO: Created: latency-svc-wc54g +Feb 25 11:20:45.975: INFO: Got endpoints: latency-svc-rwk65 [324.573294ms] +Feb 25 11:20:45.975: INFO: Got endpoints: latency-svc-2j8p4 [444.85583ms] +Feb 25 11:20:45.995: INFO: Got endpoints: latency-svc-8hr6z [464.133831ms] +Feb 25 11:20:45.996: INFO: Got endpoints: latency-svc-69jhh [414.149226ms] +Feb 25 11:20:46.028: INFO: Got endpoints: latency-svc-f5gql [301.760045ms] +Feb 25 11:20:46.029: INFO: Got endpoints: latency-svc-fm7sm [497.921371ms] +Feb 25 11:20:46.029: INFO: Got endpoints: latency-svc-47hfb [367.800143ms] +Feb 25 11:20:46.046: INFO: Got endpoints: latency-svc-mjl62 [383.831921ms] +Feb 25 11:20:46.070: INFO: Got endpoints: latency-svc-wc54g [337.680466ms] +Feb 25 11:20:46.082: INFO: Got endpoints: latency-svc-tjnxl [500.15417ms] +Feb 25 11:20:46.082: INFO: Got endpoints: latency-svc-nnq2h [463.335572ms] +Feb 25 11:20:46.083: INFO: Created: latency-svc-xwhjl +Feb 25 11:20:46.103: INFO: Got endpoints: latency-svc-ztmw7 [571.529398ms] +Feb 25 11:20:46.155: INFO: Created: latency-svc-9v47x +Feb 25 11:20:46.155: INFO: Got endpoints: latency-svc-9v47x [198.289856ms] +Feb 25 11:20:46.190: INFO: Created: latency-svc-pxfd9 +Feb 25 11:20:46.195: INFO: Got endpoints: latency-svc-xwhjl [219.981165ms] +Feb 25 11:20:46.196: INFO: Got endpoints: latency-svc-dpcg4 [500.983054ms] +Feb 25 11:20:46.196: INFO: Got endpoints: latency-svc-57bff [500.967978ms] +Feb 25 11:20:46.213: INFO: Got endpoints: latency-svc-pxfd9 [238.224937ms] +Feb 25 11:20:46.222: INFO: Created: latency-svc-zv9tf +Feb 25 11:20:46.230: INFO: Got endpoints: latency-svc-zv9tf [234.712014ms] +Feb 25 11:20:46.235: INFO: Created: latency-svc-shzqf +Feb 25 11:20:46.262: INFO: Got endpoints: latency-svc-shzqf [266.527831ms] +Feb 25 11:20:46.264: INFO: Created: latency-svc-lrdjg +Feb 25 11:20:46.292: INFO: Created: latency-svc-46sk6 +Feb 25 11:20:46.294: INFO: Got endpoints: latency-svc-lrdjg [265.876689ms] +Feb 25 11:20:46.306: INFO: Got endpoints: latency-svc-46sk6 [277.774752ms] +Feb 25 11:20:46.479: INFO: Created: latency-svc-bvwzc +Feb 25 11:20:46.480: INFO: Created: latency-svc-tnhn8 +Feb 25 11:20:46.486: INFO: Created: latency-svc-b8sqs +Feb 25 11:20:46.514: INFO: Created: latency-svc-vmn4h +Feb 25 11:20:46.514: INFO: Created: latency-svc-x96km +Feb 25 11:20:46.514: INFO: Created: latency-svc-jwv9k +Feb 25 11:20:46.515: INFO: Created: latency-svc-2vsx8 +Feb 25 11:20:46.515: INFO: Created: latency-svc-qzk66 +Feb 25 11:20:46.516: INFO: Created: latency-svc-qxn5d +Feb 25 11:20:46.517: INFO: Created: latency-svc-6z2kp +Feb 25 11:20:46.519: INFO: Created: latency-svc-hdknf +Feb 25 11:20:46.519: INFO: Created: latency-svc-zqr6p +Feb 25 11:20:46.527: INFO: Created: latency-svc-pz4sv +Feb 25 11:20:46.527: INFO: Created: latency-svc-r6ktr +Feb 25 11:20:46.527: INFO: Created: latency-svc-7xttj +Feb 25 11:20:46.542: INFO: Got endpoints: latency-svc-bvwzc [459.492615ms] +Feb 25 11:20:46.582: INFO: Got endpoints: latency-svc-2vsx8 [319.513827ms] +Feb 25 11:20:46.582: INFO: Got endpoints: latency-svc-tnhn8 [553.514728ms] +Feb 25 11:20:46.585: INFO: Got endpoints: latency-svc-qxn5d [389.166989ms] +Feb 25 11:20:46.597: INFO: Got endpoints: latency-svc-6z2kp [383.043606ms] +Feb 25 11:20:46.609: INFO: Got endpoints: latency-svc-7xttj [563.384289ms] +Feb 25 11:20:46.609: INFO: Got endpoints: latency-svc-pz4sv [527.312106ms] +Feb 25 11:20:46.628: INFO: Created: latency-svc-446f6 +Feb 25 11:20:46.670: INFO: Created: latency-svc-zxdmv +Feb 25 11:20:46.678: INFO: Created: latency-svc-nj457 +Feb 25 11:20:46.684: INFO: Got endpoints: latency-svc-hdknf [528.739945ms] +Feb 25 11:20:46.697: INFO: Got endpoints: latency-svc-zqr6p [501.970486ms] +Feb 25 11:20:46.717: INFO: Created: latency-svc-nq2kj +Feb 25 11:20:46.719: INFO: Got endpoints: latency-svc-x96km [648.721201ms] +Feb 25 11:20:46.720: INFO: Got endpoints: latency-svc-r6ktr [425.75944ms] +Feb 25 11:20:46.720: INFO: Got endpoints: latency-svc-qzk66 [524.179702ms] +Feb 25 11:20:46.721: INFO: Got endpoints: latency-svc-vmn4h [414.226701ms] +Feb 25 11:20:46.721: INFO: Got endpoints: latency-svc-b8sqs [617.964979ms] +Feb 25 11:20:46.740: INFO: Got endpoints: latency-svc-jwv9k [509.663951ms] +Feb 25 11:20:46.762: INFO: Got endpoints: latency-svc-446f6 [220.531853ms] +Feb 25 11:20:46.768: INFO: Created: latency-svc-d8ztx +Feb 25 11:20:46.784: INFO: Got endpoints: latency-svc-zxdmv [201.693337ms] +Feb 25 11:20:46.789: INFO: Got endpoints: latency-svc-nj457 [192.281607ms] +Feb 25 11:20:46.813: INFO: Got endpoints: latency-svc-nq2kj [230.071141ms] +Feb 25 11:20:46.866: INFO: Got endpoints: latency-svc-d8ztx [281.373826ms] +Feb 25 11:20:47.076: INFO: Created: latency-svc-mlrps +Feb 25 11:20:47.086: INFO: Created: latency-svc-rc2sc +Feb 25 11:20:47.087: INFO: Created: latency-svc-l6tnp +Feb 25 11:20:47.097: INFO: Created: latency-svc-xtmfb +Feb 25 11:20:47.098: INFO: Created: latency-svc-gh95s +Feb 25 11:20:47.098: INFO: Created: latency-svc-c4n62 +Feb 25 11:20:47.115: INFO: Created: latency-svc-vtm5h +Feb 25 11:20:47.115: INFO: Created: latency-svc-h2q98 +Feb 25 11:20:47.115: INFO: Created: latency-svc-qh7vd +Feb 25 11:20:47.118: INFO: Created: latency-svc-r7xk4 +Feb 25 11:20:47.119: INFO: Created: latency-svc-72dnk +Feb 25 11:20:47.123: INFO: Created: latency-svc-m47n5 +Feb 25 11:20:47.124: INFO: Created: latency-svc-gpz9r +Feb 25 11:20:47.125: INFO: Created: latency-svc-h8blc +Feb 25 11:20:47.125: INFO: Created: latency-svc-66rzs +Feb 25 11:20:47.142: INFO: Got endpoints: latency-svc-rc2sc [421.524152ms] +Feb 25 11:20:47.142: INFO: Got endpoints: latency-svc-mlrps [443.913612ms] +Feb 25 11:20:47.153: INFO: Got endpoints: latency-svc-l6tnp [468.541479ms] +Feb 25 11:20:47.156: INFO: Got endpoints: latency-svc-r7xk4 [289.733594ms] +Feb 25 11:20:47.174: INFO: Got endpoints: latency-svc-c4n62 [361.321313ms] +Feb 25 11:20:47.183: INFO: Got endpoints: latency-svc-xtmfb [462.124108ms] +Feb 25 11:20:47.193: INFO: Created: latency-svc-nv2dx +Feb 25 11:20:47.206: INFO: Got endpoints: latency-svc-vtm5h [442.647971ms] +Feb 25 11:20:47.212: INFO: Created: latency-svc-v66lk +Feb 25 11:20:47.237: INFO: Created: latency-svc-pkh86 +Feb 25 11:20:47.258: INFO: Got endpoints: latency-svc-gh95s [468.682201ms] +Feb 25 11:20:47.273: INFO: Created: latency-svc-fnlln +Feb 25 11:20:47.285: INFO: Created: latency-svc-8kbzj +Feb 25 11:20:47.311: INFO: Got endpoints: latency-svc-h2q98 [590.043309ms] +Feb 25 11:20:47.316: INFO: Created: latency-svc-fzlvf +Feb 25 11:20:47.318: INFO: Created: latency-svc-8lpb5 +Feb 25 11:20:47.340: INFO: Created: latency-svc-k8j2b +Feb 25 11:20:47.356: INFO: Created: latency-svc-2q8np +Feb 25 11:20:47.358: INFO: Got endpoints: latency-svc-qh7vd [638.626579ms] +Feb 25 11:20:47.389: INFO: Created: latency-svc-qnj6v +Feb 25 11:20:47.409: INFO: Got endpoints: latency-svc-h8blc [687.604145ms] +Feb 25 11:20:47.444: INFO: Created: latency-svc-zmgvt +Feb 25 11:20:47.464: INFO: Got endpoints: latency-svc-gpz9r [854.996255ms] +Feb 25 11:20:47.486: INFO: Created: latency-svc-7fhs7 +Feb 25 11:20:47.515: INFO: Got endpoints: latency-svc-66rzs [731.494612ms] +Feb 25 11:20:47.557: INFO: Got endpoints: latency-svc-m47n5 [947.858421ms] +Feb 25 11:20:47.581: INFO: Created: latency-svc-zhfzz +Feb 25 11:20:47.614: INFO: Got endpoints: latency-svc-72dnk [873.963924ms] +Feb 25 11:20:47.621: INFO: Created: latency-svc-gm4xd +Feb 25 11:20:47.667: INFO: Got endpoints: latency-svc-nv2dx [510.001767ms] +Feb 25 11:20:47.685: INFO: Created: latency-svc-kclx8 +Feb 25 11:20:47.711: INFO: Got endpoints: latency-svc-v66lk [558.160651ms] +Feb 25 11:20:47.715: INFO: Created: latency-svc-6nfng +Feb 25 11:20:47.775: INFO: Got endpoints: latency-svc-pkh86 [633.554587ms] +Feb 25 11:20:47.778: INFO: Created: latency-svc-8vs6f +Feb 25 11:20:47.802: INFO: Created: latency-svc-zh7kr +Feb 25 11:20:47.809: INFO: Got endpoints: latency-svc-fnlln [666.888755ms] +Feb 25 11:20:47.882: INFO: Got endpoints: latency-svc-8kbzj [708.023736ms] +Feb 25 11:20:47.920: INFO: Got endpoints: latency-svc-fzlvf [736.928668ms] +Feb 25 11:20:47.920: INFO: Created: latency-svc-wgpj2 +Feb 25 11:20:47.967: INFO: Got endpoints: latency-svc-8lpb5 [759.874548ms] +Feb 25 11:20:47.968: INFO: Created: latency-svc-ztdjq +Feb 25 11:20:47.979: INFO: Created: latency-svc-s7zfm +Feb 25 11:20:48.003: INFO: Created: latency-svc-vz9pb +Feb 25 11:20:48.012: INFO: Got endpoints: latency-svc-k8j2b [752.858978ms] +Feb 25 11:20:48.045: INFO: Created: latency-svc-vd5bs +Feb 25 11:20:48.066: INFO: Got endpoints: latency-svc-2q8np [755.454405ms] +Feb 25 11:20:48.090: INFO: Created: latency-svc-5bnx5 +Feb 25 11:20:48.113: INFO: Got endpoints: latency-svc-qnj6v [755.220786ms] +Feb 25 11:20:48.168: INFO: Got endpoints: latency-svc-zmgvt [759.649164ms] +Feb 25 11:20:48.196: INFO: Created: latency-svc-6d8n7 +Feb 25 11:20:48.217: INFO: Got endpoints: latency-svc-7fhs7 [752.348553ms] +Feb 25 11:20:48.258: INFO: Created: latency-svc-gxwxq +Feb 25 11:20:48.269: INFO: Got endpoints: latency-svc-zhfzz [753.202628ms] +Feb 25 11:20:48.289: INFO: Created: latency-svc-ljc6f +Feb 25 11:20:48.316: INFO: Got endpoints: latency-svc-gm4xd [758.811017ms] +Feb 25 11:20:48.325: INFO: Created: latency-svc-rzcsq +Feb 25 11:20:48.352: INFO: Got endpoints: latency-svc-kclx8 [738.460496ms] +Feb 25 11:20:48.375: INFO: Created: latency-svc-dst9m +Feb 25 11:20:48.398: INFO: Created: latency-svc-w5xm8 +Feb 25 11:20:48.407: INFO: Got endpoints: latency-svc-6nfng [740.59616ms] +Feb 25 11:20:48.432: INFO: Created: latency-svc-tq9sg +Feb 25 11:20:48.460: INFO: Got endpoints: latency-svc-8vs6f [748.780828ms] +Feb 25 11:20:48.485: INFO: Created: latency-svc-wtzr2 +Feb 25 11:20:48.511: INFO: Got endpoints: latency-svc-zh7kr [735.160008ms] +Feb 25 11:20:48.544: INFO: Created: latency-svc-7l7j6 +Feb 25 11:20:48.555: INFO: Got endpoints: latency-svc-wgpj2 [744.488755ms] +Feb 25 11:20:48.588: INFO: Created: latency-svc-2vl27 +Feb 25 11:20:48.612: INFO: Got endpoints: latency-svc-ztdjq [729.512291ms] +Feb 25 11:20:48.633: INFO: Created: latency-svc-gn8nn +Feb 25 11:20:48.651: INFO: Got endpoints: latency-svc-s7zfm [730.766652ms] +Feb 25 11:20:48.683: INFO: Created: latency-svc-6n674 +Feb 25 11:20:48.710: INFO: Got endpoints: latency-svc-vz9pb [743.392974ms] +Feb 25 11:20:48.728: INFO: Created: latency-svc-6nxld +Feb 25 11:20:48.760: INFO: Got endpoints: latency-svc-vd5bs [747.562171ms] +Feb 25 11:20:48.778: INFO: Created: latency-svc-k55fc +Feb 25 11:20:48.801: INFO: Got endpoints: latency-svc-5bnx5 [734.353982ms] +Feb 25 11:20:48.826: INFO: Created: latency-svc-qcd28 +Feb 25 11:20:48.890: INFO: Got endpoints: latency-svc-6d8n7 [777.032728ms] +Feb 25 11:20:48.935: INFO: Got endpoints: latency-svc-gxwxq [766.141586ms] +Feb 25 11:20:48.998: INFO: Got endpoints: latency-svc-ljc6f [781.312834ms] +Feb 25 11:20:49.011: INFO: Created: latency-svc-5kx9m +Feb 25 11:20:49.029: INFO: Got endpoints: latency-svc-rzcsq [759.843702ms] +Feb 25 11:20:49.045: INFO: Created: latency-svc-fxsjc +Feb 25 11:20:49.045: INFO: Created: latency-svc-vhjhv +Feb 25 11:20:49.055: INFO: Got endpoints: latency-svc-dst9m [738.488577ms] +Feb 25 11:20:49.113: INFO: Got endpoints: latency-svc-w5xm8 [760.345211ms] +Feb 25 11:20:49.166: INFO: Created: latency-svc-cpcq8 +Feb 25 11:20:49.171: INFO: Got endpoints: latency-svc-tq9sg [763.979765ms] +Feb 25 11:20:49.208: INFO: Got endpoints: latency-svc-wtzr2 [748.104799ms] +Feb 25 11:20:49.247: INFO: Created: latency-svc-8bc6f +Feb 25 11:20:49.313: INFO: Got endpoints: latency-svc-7l7j6 [802.209556ms] +Feb 25 11:20:49.357: INFO: Got endpoints: latency-svc-2vl27 [801.384971ms] +Feb 25 11:20:49.389: INFO: Created: latency-svc-fnzg9 +Feb 25 11:20:49.407: INFO: Got endpoints: latency-svc-gn8nn [795.309235ms] +Feb 25 11:20:49.413: INFO: Created: latency-svc-9kqst +Feb 25 11:20:49.428: INFO: Got endpoints: latency-svc-6n674 [776.956438ms] +Feb 25 11:20:49.437: INFO: Created: latency-svc-z7zjm +Feb 25 11:20:49.453: INFO: Got endpoints: latency-svc-6nxld [742.323759ms] +Feb 25 11:20:49.466: INFO: Created: latency-svc-tjcz4 +Feb 25 11:20:49.476: INFO: Created: latency-svc-hm4nn +Feb 25 11:20:49.516: INFO: Created: latency-svc-m2sbz +Feb 25 11:20:49.517: INFO: Got endpoints: latency-svc-k55fc [757.261434ms] +Feb 25 11:20:49.543: INFO: Created: latency-svc-qhf6q +Feb 25 11:20:49.573: INFO: Got endpoints: latency-svc-qcd28 [772.213281ms] +Feb 25 11:20:49.578: INFO: Created: latency-svc-sjdgq +Feb 25 11:20:49.613: INFO: Created: latency-svc-g8nh4 +Feb 25 11:20:49.617: INFO: Got endpoints: latency-svc-5kx9m [726.724858ms] +Feb 25 11:20:49.696: INFO: Created: latency-svc-4blqf +Feb 25 11:20:49.706: INFO: Got endpoints: latency-svc-vhjhv [771.387184ms] +Feb 25 11:20:49.735: INFO: Created: latency-svc-t4w6m +Feb 25 11:20:49.776: INFO: Got endpoints: latency-svc-fxsjc [777.327352ms] +Feb 25 11:20:49.780: INFO: Created: latency-svc-sdw62 +Feb 25 11:20:49.798: INFO: Got endpoints: latency-svc-cpcq8 [769.650509ms] +Feb 25 11:20:49.836: INFO: Created: latency-svc-v8g9b +Feb 25 11:20:49.842: INFO: Got endpoints: latency-svc-8bc6f [787.24503ms] +Feb 25 11:20:49.893: INFO: Got endpoints: latency-svc-fnzg9 [779.65108ms] +Feb 25 11:20:49.893: INFO: Created: latency-svc-2cv9v +Feb 25 11:20:49.931: INFO: Created: latency-svc-vs5ng +Feb 25 11:20:49.935: INFO: Got endpoints: latency-svc-9kqst [763.628966ms] +Feb 25 11:20:49.936: INFO: Created: latency-svc-r2vlh +Feb 25 11:20:49.954: INFO: Got endpoints: latency-svc-z7zjm [745.989859ms] +Feb 25 11:20:49.999: INFO: Got endpoints: latency-svc-tjcz4 [685.868755ms] +Feb 25 11:20:50.007: INFO: Created: latency-svc-vlmrm +Feb 25 11:20:50.036: INFO: Created: latency-svc-n78mk +Feb 25 11:20:50.039: INFO: Created: latency-svc-lws5j +Feb 25 11:20:50.051: INFO: Got endpoints: latency-svc-hm4nn [693.825287ms] +Feb 25 11:20:50.080: INFO: Created: latency-svc-wtp6s +Feb 25 11:20:50.103: INFO: Got endpoints: latency-svc-m2sbz [695.154883ms] +Feb 25 11:20:50.130: INFO: Created: latency-svc-4mhl4 +Feb 25 11:20:50.156: INFO: Got endpoints: latency-svc-qhf6q [728.345793ms] +Feb 25 11:20:50.187: INFO: Created: latency-svc-h9r2n +Feb 25 11:20:50.200: INFO: Got endpoints: latency-svc-sjdgq [746.709132ms] +Feb 25 11:20:50.219: INFO: Created: latency-svc-5n7cm +Feb 25 11:20:50.256: INFO: Got endpoints: latency-svc-g8nh4 [738.6959ms] +Feb 25 11:20:50.276: INFO: Created: latency-svc-prr5f +Feb 25 11:20:50.302: INFO: Got endpoints: latency-svc-4blqf [725.628793ms] +Feb 25 11:20:50.319: INFO: Created: latency-svc-42jlr +Feb 25 11:20:50.360: INFO: Got endpoints: latency-svc-t4w6m [742.57398ms] +Feb 25 11:20:50.384: INFO: Created: latency-svc-wbkp2 +Feb 25 11:20:50.401: INFO: Got endpoints: latency-svc-sdw62 [693.691925ms] +Feb 25 11:20:50.418: INFO: Created: latency-svc-h88zv +Feb 25 11:20:50.454: INFO: Got endpoints: latency-svc-v8g9b [676.987651ms] +Feb 25 11:20:50.471: INFO: Created: latency-svc-hgdrn +Feb 25 11:20:50.501: INFO: Got endpoints: latency-svc-2cv9v [698.154723ms] +Feb 25 11:20:50.519: INFO: Created: latency-svc-ssrxf +Feb 25 11:20:50.549: INFO: Got endpoints: latency-svc-vs5ng [706.630763ms] +Feb 25 11:20:50.573: INFO: Created: latency-svc-mf6bz +Feb 25 11:20:50.599: INFO: Got endpoints: latency-svc-r2vlh [705.947396ms] +Feb 25 11:20:50.629: INFO: Created: latency-svc-rbcfz +Feb 25 11:20:50.654: INFO: Got endpoints: latency-svc-vlmrm [718.745262ms] +Feb 25 11:20:50.677: INFO: Created: latency-svc-k22dt +Feb 25 11:20:50.702: INFO: Got endpoints: latency-svc-lws5j [746.917144ms] +Feb 25 11:20:50.727: INFO: Created: latency-svc-lklrf +Feb 25 11:20:50.759: INFO: Got endpoints: latency-svc-n78mk [759.632921ms] +Feb 25 11:20:50.775: INFO: Created: latency-svc-wvjrc +Feb 25 11:20:50.807: INFO: Got endpoints: latency-svc-wtp6s [751.810695ms] +Feb 25 11:20:50.836: INFO: Created: latency-svc-pnvch +Feb 25 11:20:50.850: INFO: Got endpoints: latency-svc-4mhl4 [747.42446ms] +Feb 25 11:20:50.899: INFO: Created: latency-svc-8jg4l +Feb 25 11:20:50.929: INFO: Got endpoints: latency-svc-h9r2n [772.529815ms] +Feb 25 11:20:50.959: INFO: Got endpoints: latency-svc-5n7cm [759.613512ms] +Feb 25 11:20:50.973: INFO: Created: latency-svc-lccqg +Feb 25 11:20:51.011: INFO: Got endpoints: latency-svc-prr5f [754.942955ms] +Feb 25 11:20:51.012: INFO: Created: latency-svc-zqbnz +Feb 25 11:20:51.036: INFO: Created: latency-svc-2xt9g +Feb 25 11:20:51.052: INFO: Got endpoints: latency-svc-42jlr [749.695502ms] +Feb 25 11:20:51.072: INFO: Created: latency-svc-kjw4s +Feb 25 11:20:51.103: INFO: Got endpoints: latency-svc-wbkp2 [742.662862ms] +Feb 25 11:20:51.131: INFO: Created: latency-svc-vwz5v +Feb 25 11:20:51.154: INFO: Got endpoints: latency-svc-h88zv [752.251462ms] +Feb 25 11:20:51.175: INFO: Created: latency-svc-d4vgr +Feb 25 11:20:51.201: INFO: Got endpoints: latency-svc-hgdrn [746.457775ms] +Feb 25 11:20:51.220: INFO: Created: latency-svc-q2wvn +Feb 25 11:20:51.250: INFO: Got endpoints: latency-svc-ssrxf [748.983041ms] +Feb 25 11:20:51.275: INFO: Created: latency-svc-9knk2 +Feb 25 11:20:51.298: INFO: Got endpoints: latency-svc-mf6bz [749.494956ms] +Feb 25 11:20:51.332: INFO: Created: latency-svc-xlgzp +Feb 25 11:20:51.356: INFO: Got endpoints: latency-svc-rbcfz [756.433044ms] +Feb 25 11:20:51.400: INFO: Created: latency-svc-hp4gl +Feb 25 11:20:51.416: INFO: Got endpoints: latency-svc-k22dt [761.920419ms] +Feb 25 11:20:51.473: INFO: Got endpoints: latency-svc-lklrf [771.783641ms] +Feb 25 11:20:51.492: INFO: Created: latency-svc-wclzb +Feb 25 11:20:51.537: INFO: Created: latency-svc-xnc6n +Feb 25 11:20:51.568: INFO: Got endpoints: latency-svc-wvjrc [809.196576ms] +Feb 25 11:20:51.588: INFO: Got endpoints: latency-svc-pnvch [781.160524ms] +Feb 25 11:20:51.628: INFO: Got endpoints: latency-svc-8jg4l [777.318469ms] +Feb 25 11:20:51.628: INFO: Created: latency-svc-n768j +Feb 25 11:20:51.671: INFO: Got endpoints: latency-svc-lccqg [742.089109ms] +Feb 25 11:20:51.684: INFO: Created: latency-svc-gfjjd +Feb 25 11:20:51.701: INFO: Created: latency-svc-bq2vf +Feb 25 11:20:51.718: INFO: Created: latency-svc-v8k24 +Feb 25 11:20:51.727: INFO: Got endpoints: latency-svc-zqbnz [767.246914ms] +Feb 25 11:20:51.760: INFO: Got endpoints: latency-svc-2xt9g [748.725664ms] +Feb 25 11:20:51.781: INFO: Created: latency-svc-7rc9d +Feb 25 11:20:51.803: INFO: Created: latency-svc-njwkm +Feb 25 11:20:51.810: INFO: Got endpoints: latency-svc-kjw4s [757.263385ms] +Feb 25 11:20:51.862: INFO: Got endpoints: latency-svc-vwz5v [759.472115ms] +Feb 25 11:20:51.881: INFO: Created: latency-svc-62jxp +Feb 25 11:20:51.902: INFO: Got endpoints: latency-svc-d4vgr [748.715235ms] +Feb 25 11:20:51.929: INFO: Created: latency-svc-zdk9d +Feb 25 11:20:51.948: INFO: Created: latency-svc-pbsz7 +Feb 25 11:20:51.967: INFO: Got endpoints: latency-svc-q2wvn [765.855317ms] +Feb 25 11:20:51.991: INFO: Created: latency-svc-82z5r +Feb 25 11:20:52.006: INFO: Got endpoints: latency-svc-9knk2 [755.853327ms] +Feb 25 11:20:52.038: INFO: Created: latency-svc-fdn4m +Feb 25 11:20:52.064: INFO: Got endpoints: latency-svc-xlgzp [765.60524ms] +Feb 25 11:20:52.100: INFO: Created: latency-svc-7cvnc +Feb 25 11:20:52.112: INFO: Got endpoints: latency-svc-hp4gl [756.091386ms] +Feb 25 11:20:52.157: INFO: Got endpoints: latency-svc-wclzb [741.092133ms] +Feb 25 11:20:52.158: INFO: Created: latency-svc-k7r76 +Feb 25 11:20:52.201: INFO: Got endpoints: latency-svc-xnc6n [727.204527ms] +Feb 25 11:20:52.205: INFO: Created: latency-svc-f48gp +Feb 25 11:20:52.231: INFO: Created: latency-svc-8t9f2 +Feb 25 11:20:52.259: INFO: Got endpoints: latency-svc-n768j [690.875275ms] +Feb 25 11:20:52.301: INFO: Created: latency-svc-jcg9s +Feb 25 11:20:52.307: INFO: Got endpoints: latency-svc-gfjjd [718.849847ms] +Feb 25 11:20:52.328: INFO: Created: latency-svc-jmxcd +Feb 25 11:20:52.351: INFO: Got endpoints: latency-svc-bq2vf [723.403632ms] +Feb 25 11:20:52.374: INFO: Created: latency-svc-96wqv +Feb 25 11:20:52.402: INFO: Got endpoints: latency-svc-v8k24 [730.412273ms] +Feb 25 11:20:52.422: INFO: Created: latency-svc-sqtdh +Feb 25 11:20:52.450: INFO: Got endpoints: latency-svc-7rc9d [723.312363ms] +Feb 25 11:20:52.469: INFO: Created: latency-svc-fbl46 +Feb 25 11:20:52.498: INFO: Got endpoints: latency-svc-njwkm [738.011572ms] +Feb 25 11:20:52.522: INFO: Created: latency-svc-c9cmc +Feb 25 11:20:52.558: INFO: Got endpoints: latency-svc-62jxp [748.501112ms] +Feb 25 11:20:52.574: INFO: Created: latency-svc-mj5m6 +Feb 25 11:20:52.599: INFO: Got endpoints: latency-svc-zdk9d [737.057642ms] +Feb 25 11:20:52.620: INFO: Created: latency-svc-kpndk +Feb 25 11:20:52.655: INFO: Got endpoints: latency-svc-pbsz7 [752.025709ms] +Feb 25 11:20:52.701: INFO: Got endpoints: latency-svc-82z5r [733.799554ms] +Feb 25 11:20:52.756: INFO: Got endpoints: latency-svc-fdn4m [749.445868ms] +Feb 25 11:20:52.802: INFO: Got endpoints: latency-svc-7cvnc [737.692225ms] +Feb 25 11:20:52.942: INFO: Got endpoints: latency-svc-k7r76 [829.884512ms] +Feb 25 11:20:52.943: INFO: Got endpoints: latency-svc-f48gp [784.991143ms] +Feb 25 11:20:52.970: INFO: Got endpoints: latency-svc-8t9f2 [769.406404ms] +Feb 25 11:20:53.000: INFO: Got endpoints: latency-svc-jcg9s [740.537287ms] +Feb 25 11:20:53.063: INFO: Got endpoints: latency-svc-jmxcd [755.997533ms] +Feb 25 11:20:53.124: INFO: Got endpoints: latency-svc-96wqv [772.614947ms] +Feb 25 11:20:53.157: INFO: Got endpoints: latency-svc-sqtdh [754.689275ms] +Feb 25 11:20:53.239: INFO: Got endpoints: latency-svc-fbl46 [788.353429ms] +Feb 25 11:20:53.290: INFO: Got endpoints: latency-svc-c9cmc [792.01337ms] +Feb 25 11:20:53.308: INFO: Got endpoints: latency-svc-mj5m6 [749.774431ms] +Feb 25 11:20:53.367: INFO: Got endpoints: latency-svc-kpndk [767.191474ms] +Feb 25 11:20:53.367: INFO: Latencies: [60.695301ms 91.0181ms 132.993021ms 147.591886ms 156.861217ms 188.803702ms 192.281607ms 198.289856ms 201.693337ms 202.787455ms 219.981165ms 220.531853ms 223.755658ms 230.071141ms 234.712014ms 238.224937ms 265.876689ms 266.527831ms 277.774752ms 281.373826ms 289.733594ms 300.913142ms 301.760045ms 319.513827ms 324.573294ms 330.08416ms 336.955898ms 337.680466ms 353.01572ms 361.321313ms 367.800143ms 378.509026ms 383.043606ms 383.831921ms 389.166989ms 397.483689ms 414.149226ms 414.226701ms 421.524152ms 425.75944ms 442.647971ms 443.913612ms 444.85583ms 456.128882ms 459.492615ms 462.124108ms 463.335572ms 464.133831ms 468.541479ms 468.682201ms 497.921371ms 499.948498ms 500.15417ms 500.967978ms 500.983054ms 501.970486ms 509.663951ms 510.001767ms 517.548535ms 524.179702ms 527.312106ms 528.739945ms 553.514728ms 558.160651ms 563.384289ms 571.529398ms 590.043309ms 617.964979ms 633.554587ms 638.626579ms 648.721201ms 654.40904ms 666.888755ms 670.400875ms 676.987651ms 685.868755ms 687.604145ms 690.875275ms 693.691925ms 693.825287ms 695.154883ms 698.154723ms 705.947396ms 706.630763ms 706.883466ms 708.023736ms 718.745262ms 718.849847ms 723.312363ms 723.403632ms 725.628793ms 726.724858ms 727.204527ms 728.345793ms 729.512291ms 730.412273ms 730.766652ms 731.494612ms 733.799554ms 734.353982ms 735.160008ms 736.928668ms 737.057642ms 737.692225ms 738.011572ms 738.460496ms 738.488577ms 738.6959ms 740.537287ms 740.59616ms 741.092133ms 742.089109ms 742.323759ms 742.57398ms 742.662862ms 743.392974ms 744.488755ms 745.989859ms 746.457775ms 746.709132ms 746.917144ms 747.42446ms 747.562171ms 748.104799ms 748.501112ms 748.715235ms 748.725664ms 748.780828ms 748.983041ms 749.445868ms 749.494956ms 749.695502ms 749.774431ms 751.810695ms 752.025709ms 752.251462ms 752.348553ms 752.858978ms 753.202628ms 754.689275ms 754.942955ms 755.220786ms 755.454405ms 755.853327ms 755.854578ms 755.997533ms 756.091386ms 756.433044ms 757.261434ms 757.263385ms 758.811017ms 759.472115ms 759.613512ms 759.632921ms 759.649164ms 759.843702ms 759.874548ms 760.345211ms 761.920419ms 763.628966ms 763.979765ms 765.60524ms 765.855317ms 766.141586ms 767.191474ms 767.246914ms 769.406404ms 769.650509ms 771.387184ms 771.783641ms 772.213281ms 772.529815ms 772.614947ms 776.956438ms 777.032728ms 777.318469ms 777.327352ms 779.65108ms 781.160524ms 781.312834ms 784.991143ms 787.24503ms 788.353429ms 792.01337ms 795.309235ms 796.903663ms 801.384971ms 802.209556ms 809.196576ms 810.286136ms 829.884512ms 854.284573ms 854.996255ms 862.561566ms 873.963924ms 902.843147ms 905.540065ms 915.840616ms 928.629602ms 947.858421ms] +Feb 25 11:20:53.367: INFO: 50 %ile: 735.160008ms +Feb 25 11:20:53.367: INFO: 90 %ile: 784.991143ms +Feb 25 11:20:53.367: INFO: 99 %ile: 928.629602ms +Feb 25 11:20:53.367: INFO: Total sample count: 200 +[AfterEach] [sig-network] Service endpoints latency test/e2e/framework/framework.go:187 -Dec 14 16:53:58.238: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-8437" for this suite. 12/14/22 16:53:58.243 -{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","completed":275,"skipped":5050,"failed":0} +Feb 25 11:20:53.367: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "svc-latency-4048" for this suite. 02/25/23 11:20:53.379 +{"msg":"PASSED [sig-network] Service endpoints latency should not be very high [Conformance]","completed":272,"skipped":5143,"failed":0} ------------------------------ -• [SLOW TEST] [6.439 seconds] -[sig-storage] ConfigMap -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:88 +• [SLOW TEST] [12.052 seconds] +[sig-network] Service endpoints latency +test/e2e/network/common/framework.go:23 + should not be very high [Conformance] + test/e2e/network/service_latency.go:59 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] ConfigMap + [BeforeEach] [sig-network] Service endpoints latency test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:53:51.813 - Dec 14 16:53:51.813: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename configmap 12/14/22 16:53:51.816 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:51.841 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:51.847 - [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:88 - STEP: Creating configMap with name configmap-test-volume-map-d156e7c6-a92d-4769-9bff-2e1c1a797d87 12/14/22 16:53:51.867 - STEP: Creating a pod to test consume configMaps 12/14/22 16:53:51.873 - Dec 14 16:53:51.885: INFO: Waiting up to 5m0s for pod "pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786" in namespace "configmap-8437" to be "Succeeded or Failed" - Dec 14 16:53:51.908: INFO: Pod "pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786": Phase="Pending", Reason="", readiness=false. Elapsed: 22.545311ms - Dec 14 16:53:53.919: INFO: Pod "pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786": Phase="Running", Reason="", readiness=true. Elapsed: 2.03396165s - Dec 14 16:53:55.915: INFO: Pod "pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786": Phase="Running", Reason="", readiness=false. Elapsed: 4.029530009s - Dec 14 16:53:57.915: INFO: Pod "pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.029901115s - STEP: Saw pod success 12/14/22 16:53:57.916 - Dec 14 16:53:57.916: INFO: Pod "pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786" satisfied condition "Succeeded or Failed" - Dec 14 16:53:57.923: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786 container agnhost-container: - STEP: delete the pod 12/14/22 16:53:58.217 - Dec 14 16:53:58.234: INFO: Waiting for pod pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786 to disappear - Dec 14 16:53:58.238: INFO: Pod pod-configmaps-f97e8bbe-bd6a-48f7-8a01-4d399c5df786 no longer exists - [AfterEach] [sig-storage] ConfigMap + STEP: Creating a kubernetes client 02/25/23 11:20:41.341 + Feb 25 11:20:41.341: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename svc-latency 02/25/23 11:20:41.342 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:41.371 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:41.377 + [It] should not be very high [Conformance] + test/e2e/network/service_latency.go:59 + Feb 25 11:20:41.382: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: creating replication controller svc-latency-rc in namespace svc-latency-4048 02/25/23 11:20:41.384 + I0225 11:20:41.409426 13 runners.go:193] Created replication controller with name: svc-latency-rc, namespace: svc-latency-4048, replica count: 1 + I0225 11:20:42.483420 13 runners.go:193] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + I0225 11:20:43.483760 13 runners.go:193] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + I0225 11:20:44.484295 13 runners.go:193] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + Feb 25 11:20:44.608: INFO: Created: latency-svc-gd5mr + Feb 25 11:20:44.673: INFO: Got endpoints: latency-svc-gd5mr [88.42354ms] + Feb 25 11:20:44.723: INFO: Created: latency-svc-vbjtm + Feb 25 11:20:44.729: INFO: Created: latency-svc-dthdf + Feb 25 11:20:44.735: INFO: Got endpoints: latency-svc-dthdf [60.695301ms] + Feb 25 11:20:44.762: INFO: Created: latency-svc-dwkqx + Feb 25 11:20:44.766: INFO: Got endpoints: latency-svc-vbjtm [91.0181ms] + Feb 25 11:20:44.789: INFO: Created: latency-svc-xxwmb + Feb 25 11:20:44.808: INFO: Got endpoints: latency-svc-dwkqx [132.993021ms] + Feb 25 11:20:44.820: INFO: Created: latency-svc-v2lwj + Feb 25 11:20:44.823: INFO: Got endpoints: latency-svc-xxwmb [147.591886ms] + Feb 25 11:20:44.832: INFO: Got endpoints: latency-svc-v2lwj [156.861217ms] + Feb 25 11:20:44.837: INFO: Created: latency-svc-7xtlf + Feb 25 11:20:44.865: INFO: Got endpoints: latency-svc-7xtlf [188.803702ms] + Feb 25 11:20:44.922: INFO: Created: latency-svc-4t8lk + Feb 25 11:20:44.956: INFO: Created: latency-svc-gbxlz + Feb 25 11:20:44.977: INFO: Got endpoints: latency-svc-4t8lk [300.913142ms] + Feb 25 11:20:44.990: INFO: Created: latency-svc-7mpmp + Feb 25 11:20:45.006: INFO: Got endpoints: latency-svc-gbxlz [330.08416ms] + Feb 25 11:20:45.055: INFO: Got endpoints: latency-svc-7mpmp [378.509026ms] + Feb 25 11:20:45.085: INFO: Created: latency-svc-gtqj5 + Feb 25 11:20:45.133: INFO: Got endpoints: latency-svc-gtqj5 [456.128882ms] + Feb 25 11:20:45.153: INFO: Created: latency-svc-vmbmh + Feb 25 11:20:45.176: INFO: Created: latency-svc-tjxct + Feb 25 11:20:45.176: INFO: Got endpoints: latency-svc-tjxct [499.948498ms] + Feb 25 11:20:45.194: INFO: Got endpoints: latency-svc-vmbmh [517.548535ms] + Feb 25 11:20:45.438: INFO: Created: latency-svc-9662q + Feb 25 11:20:45.438: INFO: Created: latency-svc-dmgxp + Feb 25 11:20:45.439: INFO: Created: latency-svc-pbpbq + Feb 25 11:20:45.439: INFO: Created: latency-svc-9c2nf + Feb 25 11:20:45.439: INFO: Created: latency-svc-dkhfn + Feb 25 11:20:45.440: INFO: Created: latency-svc-kjvsl + Feb 25 11:20:45.440: INFO: Created: latency-svc-2r4kx + Feb 25 11:20:45.454: INFO: Created: latency-svc-gt48c + Feb 25 11:20:45.454: INFO: Created: latency-svc-csr4k + Feb 25 11:20:45.454: INFO: Created: latency-svc-kqvgc + Feb 25 11:20:45.516: INFO: Created: latency-svc-fxpbs + Feb 25 11:20:45.517: INFO: Created: latency-svc-wjm5p + Feb 25 11:20:45.517: INFO: Created: latency-svc-5v2s9 + Feb 25 11:20:45.530: INFO: Got endpoints: latency-svc-9c2nf [353.01572ms] + Feb 25 11:20:45.530: INFO: Got endpoints: latency-svc-kjvsl [706.883466ms] + Feb 25 11:20:45.531: INFO: Got endpoints: latency-svc-dkhfn [397.483689ms] + Feb 25 11:20:45.531: INFO: Got endpoints: latency-svc-dmgxp [854.284573ms] + Feb 25 11:20:45.531: INFO: Got endpoints: latency-svc-pbpbq [336.955898ms] + Feb 25 11:20:45.554: INFO: Created: latency-svc-xdvss + Feb 25 11:20:45.555: INFO: Created: latency-svc-nqcc4 + Feb 25 11:20:45.581: INFO: Got endpoints: latency-svc-2r4kx [902.843147ms] + Feb 25 11:20:45.582: INFO: Got endpoints: latency-svc-gt48c [905.540065ms] + Feb 25 11:20:45.619: INFO: Got endpoints: latency-svc-csr4k [810.286136ms] + Feb 25 11:20:45.651: INFO: Got endpoints: latency-svc-9662q [915.840616ms] + Feb 25 11:20:45.661: INFO: Got endpoints: latency-svc-kqvgc [654.40904ms] + Feb 25 11:20:45.662: INFO: Got endpoints: latency-svc-wjm5p [796.903663ms] + Feb 25 11:20:45.695: INFO: Got endpoints: latency-svc-5v2s9 [928.629602ms] + Feb 25 11:20:45.695: INFO: Got endpoints: latency-svc-fxpbs [862.561566ms] + Feb 25 11:20:45.699: INFO: Created: latency-svc-872bc + Feb 25 11:20:45.726: INFO: Got endpoints: latency-svc-nqcc4 [670.400875ms] + Feb 25 11:20:45.733: INFO: Got endpoints: latency-svc-872bc [202.787455ms] + Feb 25 11:20:45.733: INFO: Got endpoints: latency-svc-xdvss [755.854578ms] + Feb 25 11:20:45.907: INFO: Created: latency-svc-wctmd + Feb 25 11:20:45.921: INFO: Created: latency-svc-8hr6z + Feb 25 11:20:45.923: INFO: Created: latency-svc-rwk65 + Feb 25 11:20:45.926: INFO: Created: latency-svc-nnq2h + Feb 25 11:20:45.926: INFO: Created: latency-svc-mjl62 + Feb 25 11:20:45.957: INFO: Got endpoints: latency-svc-wctmd [223.755658ms] + Feb 25 11:20:45.968: INFO: Created: latency-svc-f5gql + Feb 25 11:20:45.968: INFO: Created: latency-svc-47hfb + Feb 25 11:20:45.969: INFO: Created: latency-svc-69jhh + Feb 25 11:20:45.969: INFO: Created: latency-svc-fm7sm + Feb 25 11:20:45.969: INFO: Created: latency-svc-2j8p4 + Feb 25 11:20:45.974: INFO: Created: latency-svc-57bff + Feb 25 11:20:45.974: INFO: Created: latency-svc-dpcg4 + Feb 25 11:20:45.974: INFO: Created: latency-svc-ztmw7 + Feb 25 11:20:45.975: INFO: Created: latency-svc-tjnxl + Feb 25 11:20:45.975: INFO: Created: latency-svc-wc54g + Feb 25 11:20:45.975: INFO: Got endpoints: latency-svc-rwk65 [324.573294ms] + Feb 25 11:20:45.975: INFO: Got endpoints: latency-svc-2j8p4 [444.85583ms] + Feb 25 11:20:45.995: INFO: Got endpoints: latency-svc-8hr6z [464.133831ms] + Feb 25 11:20:45.996: INFO: Got endpoints: latency-svc-69jhh [414.149226ms] + Feb 25 11:20:46.028: INFO: Got endpoints: latency-svc-f5gql [301.760045ms] + Feb 25 11:20:46.029: INFO: Got endpoints: latency-svc-fm7sm [497.921371ms] + Feb 25 11:20:46.029: INFO: Got endpoints: latency-svc-47hfb [367.800143ms] + Feb 25 11:20:46.046: INFO: Got endpoints: latency-svc-mjl62 [383.831921ms] + Feb 25 11:20:46.070: INFO: Got endpoints: latency-svc-wc54g [337.680466ms] + Feb 25 11:20:46.082: INFO: Got endpoints: latency-svc-tjnxl [500.15417ms] + Feb 25 11:20:46.082: INFO: Got endpoints: latency-svc-nnq2h [463.335572ms] + Feb 25 11:20:46.083: INFO: Created: latency-svc-xwhjl + Feb 25 11:20:46.103: INFO: Got endpoints: latency-svc-ztmw7 [571.529398ms] + Feb 25 11:20:46.155: INFO: Created: latency-svc-9v47x + Feb 25 11:20:46.155: INFO: Got endpoints: latency-svc-9v47x [198.289856ms] + Feb 25 11:20:46.190: INFO: Created: latency-svc-pxfd9 + Feb 25 11:20:46.195: INFO: Got endpoints: latency-svc-xwhjl [219.981165ms] + Feb 25 11:20:46.196: INFO: Got endpoints: latency-svc-dpcg4 [500.983054ms] + Feb 25 11:20:46.196: INFO: Got endpoints: latency-svc-57bff [500.967978ms] + Feb 25 11:20:46.213: INFO: Got endpoints: latency-svc-pxfd9 [238.224937ms] + Feb 25 11:20:46.222: INFO: Created: latency-svc-zv9tf + Feb 25 11:20:46.230: INFO: Got endpoints: latency-svc-zv9tf [234.712014ms] + Feb 25 11:20:46.235: INFO: Created: latency-svc-shzqf + Feb 25 11:20:46.262: INFO: Got endpoints: latency-svc-shzqf [266.527831ms] + Feb 25 11:20:46.264: INFO: Created: latency-svc-lrdjg + Feb 25 11:20:46.292: INFO: Created: latency-svc-46sk6 + Feb 25 11:20:46.294: INFO: Got endpoints: latency-svc-lrdjg [265.876689ms] + Feb 25 11:20:46.306: INFO: Got endpoints: latency-svc-46sk6 [277.774752ms] + Feb 25 11:20:46.479: INFO: Created: latency-svc-bvwzc + Feb 25 11:20:46.480: INFO: Created: latency-svc-tnhn8 + Feb 25 11:20:46.486: INFO: Created: latency-svc-b8sqs + Feb 25 11:20:46.514: INFO: Created: latency-svc-vmn4h + Feb 25 11:20:46.514: INFO: Created: latency-svc-x96km + Feb 25 11:20:46.514: INFO: Created: latency-svc-jwv9k + Feb 25 11:20:46.515: INFO: Created: latency-svc-2vsx8 + Feb 25 11:20:46.515: INFO: Created: latency-svc-qzk66 + Feb 25 11:20:46.516: INFO: Created: latency-svc-qxn5d + Feb 25 11:20:46.517: INFO: Created: latency-svc-6z2kp + Feb 25 11:20:46.519: INFO: Created: latency-svc-hdknf + Feb 25 11:20:46.519: INFO: Created: latency-svc-zqr6p + Feb 25 11:20:46.527: INFO: Created: latency-svc-pz4sv + Feb 25 11:20:46.527: INFO: Created: latency-svc-r6ktr + Feb 25 11:20:46.527: INFO: Created: latency-svc-7xttj + Feb 25 11:20:46.542: INFO: Got endpoints: latency-svc-bvwzc [459.492615ms] + Feb 25 11:20:46.582: INFO: Got endpoints: latency-svc-2vsx8 [319.513827ms] + Feb 25 11:20:46.582: INFO: Got endpoints: latency-svc-tnhn8 [553.514728ms] + Feb 25 11:20:46.585: INFO: Got endpoints: latency-svc-qxn5d [389.166989ms] + Feb 25 11:20:46.597: INFO: Got endpoints: latency-svc-6z2kp [383.043606ms] + Feb 25 11:20:46.609: INFO: Got endpoints: latency-svc-7xttj [563.384289ms] + Feb 25 11:20:46.609: INFO: Got endpoints: latency-svc-pz4sv [527.312106ms] + Feb 25 11:20:46.628: INFO: Created: latency-svc-446f6 + Feb 25 11:20:46.670: INFO: Created: latency-svc-zxdmv + Feb 25 11:20:46.678: INFO: Created: latency-svc-nj457 + Feb 25 11:20:46.684: INFO: Got endpoints: latency-svc-hdknf [528.739945ms] + Feb 25 11:20:46.697: INFO: Got endpoints: latency-svc-zqr6p [501.970486ms] + Feb 25 11:20:46.717: INFO: Created: latency-svc-nq2kj + Feb 25 11:20:46.719: INFO: Got endpoints: latency-svc-x96km [648.721201ms] + Feb 25 11:20:46.720: INFO: Got endpoints: latency-svc-r6ktr [425.75944ms] + Feb 25 11:20:46.720: INFO: Got endpoints: latency-svc-qzk66 [524.179702ms] + Feb 25 11:20:46.721: INFO: Got endpoints: latency-svc-vmn4h [414.226701ms] + Feb 25 11:20:46.721: INFO: Got endpoints: latency-svc-b8sqs [617.964979ms] + Feb 25 11:20:46.740: INFO: Got endpoints: latency-svc-jwv9k [509.663951ms] + Feb 25 11:20:46.762: INFO: Got endpoints: latency-svc-446f6 [220.531853ms] + Feb 25 11:20:46.768: INFO: Created: latency-svc-d8ztx + Feb 25 11:20:46.784: INFO: Got endpoints: latency-svc-zxdmv [201.693337ms] + Feb 25 11:20:46.789: INFO: Got endpoints: latency-svc-nj457 [192.281607ms] + Feb 25 11:20:46.813: INFO: Got endpoints: latency-svc-nq2kj [230.071141ms] + Feb 25 11:20:46.866: INFO: Got endpoints: latency-svc-d8ztx [281.373826ms] + Feb 25 11:20:47.076: INFO: Created: latency-svc-mlrps + Feb 25 11:20:47.086: INFO: Created: latency-svc-rc2sc + Feb 25 11:20:47.087: INFO: Created: latency-svc-l6tnp + Feb 25 11:20:47.097: INFO: Created: latency-svc-xtmfb + Feb 25 11:20:47.098: INFO: Created: latency-svc-gh95s + Feb 25 11:20:47.098: INFO: Created: latency-svc-c4n62 + Feb 25 11:20:47.115: INFO: Created: latency-svc-vtm5h + Feb 25 11:20:47.115: INFO: Created: latency-svc-h2q98 + Feb 25 11:20:47.115: INFO: Created: latency-svc-qh7vd + Feb 25 11:20:47.118: INFO: Created: latency-svc-r7xk4 + Feb 25 11:20:47.119: INFO: Created: latency-svc-72dnk + Feb 25 11:20:47.123: INFO: Created: latency-svc-m47n5 + Feb 25 11:20:47.124: INFO: Created: latency-svc-gpz9r + Feb 25 11:20:47.125: INFO: Created: latency-svc-h8blc + Feb 25 11:20:47.125: INFO: Created: latency-svc-66rzs + Feb 25 11:20:47.142: INFO: Got endpoints: latency-svc-rc2sc [421.524152ms] + Feb 25 11:20:47.142: INFO: Got endpoints: latency-svc-mlrps [443.913612ms] + Feb 25 11:20:47.153: INFO: Got endpoints: latency-svc-l6tnp [468.541479ms] + Feb 25 11:20:47.156: INFO: Got endpoints: latency-svc-r7xk4 [289.733594ms] + Feb 25 11:20:47.174: INFO: Got endpoints: latency-svc-c4n62 [361.321313ms] + Feb 25 11:20:47.183: INFO: Got endpoints: latency-svc-xtmfb [462.124108ms] + Feb 25 11:20:47.193: INFO: Created: latency-svc-nv2dx + Feb 25 11:20:47.206: INFO: Got endpoints: latency-svc-vtm5h [442.647971ms] + Feb 25 11:20:47.212: INFO: Created: latency-svc-v66lk + Feb 25 11:20:47.237: INFO: Created: latency-svc-pkh86 + Feb 25 11:20:47.258: INFO: Got endpoints: latency-svc-gh95s [468.682201ms] + Feb 25 11:20:47.273: INFO: Created: latency-svc-fnlln + Feb 25 11:20:47.285: INFO: Created: latency-svc-8kbzj + Feb 25 11:20:47.311: INFO: Got endpoints: latency-svc-h2q98 [590.043309ms] + Feb 25 11:20:47.316: INFO: Created: latency-svc-fzlvf + Feb 25 11:20:47.318: INFO: Created: latency-svc-8lpb5 + Feb 25 11:20:47.340: INFO: Created: latency-svc-k8j2b + Feb 25 11:20:47.356: INFO: Created: latency-svc-2q8np + Feb 25 11:20:47.358: INFO: Got endpoints: latency-svc-qh7vd [638.626579ms] + Feb 25 11:20:47.389: INFO: Created: latency-svc-qnj6v + Feb 25 11:20:47.409: INFO: Got endpoints: latency-svc-h8blc [687.604145ms] + Feb 25 11:20:47.444: INFO: Created: latency-svc-zmgvt + Feb 25 11:20:47.464: INFO: Got endpoints: latency-svc-gpz9r [854.996255ms] + Feb 25 11:20:47.486: INFO: Created: latency-svc-7fhs7 + Feb 25 11:20:47.515: INFO: Got endpoints: latency-svc-66rzs [731.494612ms] + Feb 25 11:20:47.557: INFO: Got endpoints: latency-svc-m47n5 [947.858421ms] + Feb 25 11:20:47.581: INFO: Created: latency-svc-zhfzz + Feb 25 11:20:47.614: INFO: Got endpoints: latency-svc-72dnk [873.963924ms] + Feb 25 11:20:47.621: INFO: Created: latency-svc-gm4xd + Feb 25 11:20:47.667: INFO: Got endpoints: latency-svc-nv2dx [510.001767ms] + Feb 25 11:20:47.685: INFO: Created: latency-svc-kclx8 + Feb 25 11:20:47.711: INFO: Got endpoints: latency-svc-v66lk [558.160651ms] + Feb 25 11:20:47.715: INFO: Created: latency-svc-6nfng + Feb 25 11:20:47.775: INFO: Got endpoints: latency-svc-pkh86 [633.554587ms] + Feb 25 11:20:47.778: INFO: Created: latency-svc-8vs6f + Feb 25 11:20:47.802: INFO: Created: latency-svc-zh7kr + Feb 25 11:20:47.809: INFO: Got endpoints: latency-svc-fnlln [666.888755ms] + Feb 25 11:20:47.882: INFO: Got endpoints: latency-svc-8kbzj [708.023736ms] + Feb 25 11:20:47.920: INFO: Got endpoints: latency-svc-fzlvf [736.928668ms] + Feb 25 11:20:47.920: INFO: Created: latency-svc-wgpj2 + Feb 25 11:20:47.967: INFO: Got endpoints: latency-svc-8lpb5 [759.874548ms] + Feb 25 11:20:47.968: INFO: Created: latency-svc-ztdjq + Feb 25 11:20:47.979: INFO: Created: latency-svc-s7zfm + Feb 25 11:20:48.003: INFO: Created: latency-svc-vz9pb + Feb 25 11:20:48.012: INFO: Got endpoints: latency-svc-k8j2b [752.858978ms] + Feb 25 11:20:48.045: INFO: Created: latency-svc-vd5bs + Feb 25 11:20:48.066: INFO: Got endpoints: latency-svc-2q8np [755.454405ms] + Feb 25 11:20:48.090: INFO: Created: latency-svc-5bnx5 + Feb 25 11:20:48.113: INFO: Got endpoints: latency-svc-qnj6v [755.220786ms] + Feb 25 11:20:48.168: INFO: Got endpoints: latency-svc-zmgvt [759.649164ms] + Feb 25 11:20:48.196: INFO: Created: latency-svc-6d8n7 + Feb 25 11:20:48.217: INFO: Got endpoints: latency-svc-7fhs7 [752.348553ms] + Feb 25 11:20:48.258: INFO: Created: latency-svc-gxwxq + Feb 25 11:20:48.269: INFO: Got endpoints: latency-svc-zhfzz [753.202628ms] + Feb 25 11:20:48.289: INFO: Created: latency-svc-ljc6f + Feb 25 11:20:48.316: INFO: Got endpoints: latency-svc-gm4xd [758.811017ms] + Feb 25 11:20:48.325: INFO: Created: latency-svc-rzcsq + Feb 25 11:20:48.352: INFO: Got endpoints: latency-svc-kclx8 [738.460496ms] + Feb 25 11:20:48.375: INFO: Created: latency-svc-dst9m + Feb 25 11:20:48.398: INFO: Created: latency-svc-w5xm8 + Feb 25 11:20:48.407: INFO: Got endpoints: latency-svc-6nfng [740.59616ms] + Feb 25 11:20:48.432: INFO: Created: latency-svc-tq9sg + Feb 25 11:20:48.460: INFO: Got endpoints: latency-svc-8vs6f [748.780828ms] + Feb 25 11:20:48.485: INFO: Created: latency-svc-wtzr2 + Feb 25 11:20:48.511: INFO: Got endpoints: latency-svc-zh7kr [735.160008ms] + Feb 25 11:20:48.544: INFO: Created: latency-svc-7l7j6 + Feb 25 11:20:48.555: INFO: Got endpoints: latency-svc-wgpj2 [744.488755ms] + Feb 25 11:20:48.588: INFO: Created: latency-svc-2vl27 + Feb 25 11:20:48.612: INFO: Got endpoints: latency-svc-ztdjq [729.512291ms] + Feb 25 11:20:48.633: INFO: Created: latency-svc-gn8nn + Feb 25 11:20:48.651: INFO: Got endpoints: latency-svc-s7zfm [730.766652ms] + Feb 25 11:20:48.683: INFO: Created: latency-svc-6n674 + Feb 25 11:20:48.710: INFO: Got endpoints: latency-svc-vz9pb [743.392974ms] + Feb 25 11:20:48.728: INFO: Created: latency-svc-6nxld + Feb 25 11:20:48.760: INFO: Got endpoints: latency-svc-vd5bs [747.562171ms] + Feb 25 11:20:48.778: INFO: Created: latency-svc-k55fc + Feb 25 11:20:48.801: INFO: Got endpoints: latency-svc-5bnx5 [734.353982ms] + Feb 25 11:20:48.826: INFO: Created: latency-svc-qcd28 + Feb 25 11:20:48.890: INFO: Got endpoints: latency-svc-6d8n7 [777.032728ms] + Feb 25 11:20:48.935: INFO: Got endpoints: latency-svc-gxwxq [766.141586ms] + Feb 25 11:20:48.998: INFO: Got endpoints: latency-svc-ljc6f [781.312834ms] + Feb 25 11:20:49.011: INFO: Created: latency-svc-5kx9m + Feb 25 11:20:49.029: INFO: Got endpoints: latency-svc-rzcsq [759.843702ms] + Feb 25 11:20:49.045: INFO: Created: latency-svc-fxsjc + Feb 25 11:20:49.045: INFO: Created: latency-svc-vhjhv + Feb 25 11:20:49.055: INFO: Got endpoints: latency-svc-dst9m [738.488577ms] + Feb 25 11:20:49.113: INFO: Got endpoints: latency-svc-w5xm8 [760.345211ms] + Feb 25 11:20:49.166: INFO: Created: latency-svc-cpcq8 + Feb 25 11:20:49.171: INFO: Got endpoints: latency-svc-tq9sg [763.979765ms] + Feb 25 11:20:49.208: INFO: Got endpoints: latency-svc-wtzr2 [748.104799ms] + Feb 25 11:20:49.247: INFO: Created: latency-svc-8bc6f + Feb 25 11:20:49.313: INFO: Got endpoints: latency-svc-7l7j6 [802.209556ms] + Feb 25 11:20:49.357: INFO: Got endpoints: latency-svc-2vl27 [801.384971ms] + Feb 25 11:20:49.389: INFO: Created: latency-svc-fnzg9 + Feb 25 11:20:49.407: INFO: Got endpoints: latency-svc-gn8nn [795.309235ms] + Feb 25 11:20:49.413: INFO: Created: latency-svc-9kqst + Feb 25 11:20:49.428: INFO: Got endpoints: latency-svc-6n674 [776.956438ms] + Feb 25 11:20:49.437: INFO: Created: latency-svc-z7zjm + Feb 25 11:20:49.453: INFO: Got endpoints: latency-svc-6nxld [742.323759ms] + Feb 25 11:20:49.466: INFO: Created: latency-svc-tjcz4 + Feb 25 11:20:49.476: INFO: Created: latency-svc-hm4nn + Feb 25 11:20:49.516: INFO: Created: latency-svc-m2sbz + Feb 25 11:20:49.517: INFO: Got endpoints: latency-svc-k55fc [757.261434ms] + Feb 25 11:20:49.543: INFO: Created: latency-svc-qhf6q + Feb 25 11:20:49.573: INFO: Got endpoints: latency-svc-qcd28 [772.213281ms] + Feb 25 11:20:49.578: INFO: Created: latency-svc-sjdgq + Feb 25 11:20:49.613: INFO: Created: latency-svc-g8nh4 + Feb 25 11:20:49.617: INFO: Got endpoints: latency-svc-5kx9m [726.724858ms] + Feb 25 11:20:49.696: INFO: Created: latency-svc-4blqf + Feb 25 11:20:49.706: INFO: Got endpoints: latency-svc-vhjhv [771.387184ms] + Feb 25 11:20:49.735: INFO: Created: latency-svc-t4w6m + Feb 25 11:20:49.776: INFO: Got endpoints: latency-svc-fxsjc [777.327352ms] + Feb 25 11:20:49.780: INFO: Created: latency-svc-sdw62 + Feb 25 11:20:49.798: INFO: Got endpoints: latency-svc-cpcq8 [769.650509ms] + Feb 25 11:20:49.836: INFO: Created: latency-svc-v8g9b + Feb 25 11:20:49.842: INFO: Got endpoints: latency-svc-8bc6f [787.24503ms] + Feb 25 11:20:49.893: INFO: Got endpoints: latency-svc-fnzg9 [779.65108ms] + Feb 25 11:20:49.893: INFO: Created: latency-svc-2cv9v + Feb 25 11:20:49.931: INFO: Created: latency-svc-vs5ng + Feb 25 11:20:49.935: INFO: Got endpoints: latency-svc-9kqst [763.628966ms] + Feb 25 11:20:49.936: INFO: Created: latency-svc-r2vlh + Feb 25 11:20:49.954: INFO: Got endpoints: latency-svc-z7zjm [745.989859ms] + Feb 25 11:20:49.999: INFO: Got endpoints: latency-svc-tjcz4 [685.868755ms] + Feb 25 11:20:50.007: INFO: Created: latency-svc-vlmrm + Feb 25 11:20:50.036: INFO: Created: latency-svc-n78mk + Feb 25 11:20:50.039: INFO: Created: latency-svc-lws5j + Feb 25 11:20:50.051: INFO: Got endpoints: latency-svc-hm4nn [693.825287ms] + Feb 25 11:20:50.080: INFO: Created: latency-svc-wtp6s + Feb 25 11:20:50.103: INFO: Got endpoints: latency-svc-m2sbz [695.154883ms] + Feb 25 11:20:50.130: INFO: Created: latency-svc-4mhl4 + Feb 25 11:20:50.156: INFO: Got endpoints: latency-svc-qhf6q [728.345793ms] + Feb 25 11:20:50.187: INFO: Created: latency-svc-h9r2n + Feb 25 11:20:50.200: INFO: Got endpoints: latency-svc-sjdgq [746.709132ms] + Feb 25 11:20:50.219: INFO: Created: latency-svc-5n7cm + Feb 25 11:20:50.256: INFO: Got endpoints: latency-svc-g8nh4 [738.6959ms] + Feb 25 11:20:50.276: INFO: Created: latency-svc-prr5f + Feb 25 11:20:50.302: INFO: Got endpoints: latency-svc-4blqf [725.628793ms] + Feb 25 11:20:50.319: INFO: Created: latency-svc-42jlr + Feb 25 11:20:50.360: INFO: Got endpoints: latency-svc-t4w6m [742.57398ms] + Feb 25 11:20:50.384: INFO: Created: latency-svc-wbkp2 + Feb 25 11:20:50.401: INFO: Got endpoints: latency-svc-sdw62 [693.691925ms] + Feb 25 11:20:50.418: INFO: Created: latency-svc-h88zv + Feb 25 11:20:50.454: INFO: Got endpoints: latency-svc-v8g9b [676.987651ms] + Feb 25 11:20:50.471: INFO: Created: latency-svc-hgdrn + Feb 25 11:20:50.501: INFO: Got endpoints: latency-svc-2cv9v [698.154723ms] + Feb 25 11:20:50.519: INFO: Created: latency-svc-ssrxf + Feb 25 11:20:50.549: INFO: Got endpoints: latency-svc-vs5ng [706.630763ms] + Feb 25 11:20:50.573: INFO: Created: latency-svc-mf6bz + Feb 25 11:20:50.599: INFO: Got endpoints: latency-svc-r2vlh [705.947396ms] + Feb 25 11:20:50.629: INFO: Created: latency-svc-rbcfz + Feb 25 11:20:50.654: INFO: Got endpoints: latency-svc-vlmrm [718.745262ms] + Feb 25 11:20:50.677: INFO: Created: latency-svc-k22dt + Feb 25 11:20:50.702: INFO: Got endpoints: latency-svc-lws5j [746.917144ms] + Feb 25 11:20:50.727: INFO: Created: latency-svc-lklrf + Feb 25 11:20:50.759: INFO: Got endpoints: latency-svc-n78mk [759.632921ms] + Feb 25 11:20:50.775: INFO: Created: latency-svc-wvjrc + Feb 25 11:20:50.807: INFO: Got endpoints: latency-svc-wtp6s [751.810695ms] + Feb 25 11:20:50.836: INFO: Created: latency-svc-pnvch + Feb 25 11:20:50.850: INFO: Got endpoints: latency-svc-4mhl4 [747.42446ms] + Feb 25 11:20:50.899: INFO: Created: latency-svc-8jg4l + Feb 25 11:20:50.929: INFO: Got endpoints: latency-svc-h9r2n [772.529815ms] + Feb 25 11:20:50.959: INFO: Got endpoints: latency-svc-5n7cm [759.613512ms] + Feb 25 11:20:50.973: INFO: Created: latency-svc-lccqg + Feb 25 11:20:51.011: INFO: Got endpoints: latency-svc-prr5f [754.942955ms] + Feb 25 11:20:51.012: INFO: Created: latency-svc-zqbnz + Feb 25 11:20:51.036: INFO: Created: latency-svc-2xt9g + Feb 25 11:20:51.052: INFO: Got endpoints: latency-svc-42jlr [749.695502ms] + Feb 25 11:20:51.072: INFO: Created: latency-svc-kjw4s + Feb 25 11:20:51.103: INFO: Got endpoints: latency-svc-wbkp2 [742.662862ms] + Feb 25 11:20:51.131: INFO: Created: latency-svc-vwz5v + Feb 25 11:20:51.154: INFO: Got endpoints: latency-svc-h88zv [752.251462ms] + Feb 25 11:20:51.175: INFO: Created: latency-svc-d4vgr + Feb 25 11:20:51.201: INFO: Got endpoints: latency-svc-hgdrn [746.457775ms] + Feb 25 11:20:51.220: INFO: Created: latency-svc-q2wvn + Feb 25 11:20:51.250: INFO: Got endpoints: latency-svc-ssrxf [748.983041ms] + Feb 25 11:20:51.275: INFO: Created: latency-svc-9knk2 + Feb 25 11:20:51.298: INFO: Got endpoints: latency-svc-mf6bz [749.494956ms] + Feb 25 11:20:51.332: INFO: Created: latency-svc-xlgzp + Feb 25 11:20:51.356: INFO: Got endpoints: latency-svc-rbcfz [756.433044ms] + Feb 25 11:20:51.400: INFO: Created: latency-svc-hp4gl + Feb 25 11:20:51.416: INFO: Got endpoints: latency-svc-k22dt [761.920419ms] + Feb 25 11:20:51.473: INFO: Got endpoints: latency-svc-lklrf [771.783641ms] + Feb 25 11:20:51.492: INFO: Created: latency-svc-wclzb + Feb 25 11:20:51.537: INFO: Created: latency-svc-xnc6n + Feb 25 11:20:51.568: INFO: Got endpoints: latency-svc-wvjrc [809.196576ms] + Feb 25 11:20:51.588: INFO: Got endpoints: latency-svc-pnvch [781.160524ms] + Feb 25 11:20:51.628: INFO: Got endpoints: latency-svc-8jg4l [777.318469ms] + Feb 25 11:20:51.628: INFO: Created: latency-svc-n768j + Feb 25 11:20:51.671: INFO: Got endpoints: latency-svc-lccqg [742.089109ms] + Feb 25 11:20:51.684: INFO: Created: latency-svc-gfjjd + Feb 25 11:20:51.701: INFO: Created: latency-svc-bq2vf + Feb 25 11:20:51.718: INFO: Created: latency-svc-v8k24 + Feb 25 11:20:51.727: INFO: Got endpoints: latency-svc-zqbnz [767.246914ms] + Feb 25 11:20:51.760: INFO: Got endpoints: latency-svc-2xt9g [748.725664ms] + Feb 25 11:20:51.781: INFO: Created: latency-svc-7rc9d + Feb 25 11:20:51.803: INFO: Created: latency-svc-njwkm + Feb 25 11:20:51.810: INFO: Got endpoints: latency-svc-kjw4s [757.263385ms] + Feb 25 11:20:51.862: INFO: Got endpoints: latency-svc-vwz5v [759.472115ms] + Feb 25 11:20:51.881: INFO: Created: latency-svc-62jxp + Feb 25 11:20:51.902: INFO: Got endpoints: latency-svc-d4vgr [748.715235ms] + Feb 25 11:20:51.929: INFO: Created: latency-svc-zdk9d + Feb 25 11:20:51.948: INFO: Created: latency-svc-pbsz7 + Feb 25 11:20:51.967: INFO: Got endpoints: latency-svc-q2wvn [765.855317ms] + Feb 25 11:20:51.991: INFO: Created: latency-svc-82z5r + Feb 25 11:20:52.006: INFO: Got endpoints: latency-svc-9knk2 [755.853327ms] + Feb 25 11:20:52.038: INFO: Created: latency-svc-fdn4m + Feb 25 11:20:52.064: INFO: Got endpoints: latency-svc-xlgzp [765.60524ms] + Feb 25 11:20:52.100: INFO: Created: latency-svc-7cvnc + Feb 25 11:20:52.112: INFO: Got endpoints: latency-svc-hp4gl [756.091386ms] + Feb 25 11:20:52.157: INFO: Got endpoints: latency-svc-wclzb [741.092133ms] + Feb 25 11:20:52.158: INFO: Created: latency-svc-k7r76 + Feb 25 11:20:52.201: INFO: Got endpoints: latency-svc-xnc6n [727.204527ms] + Feb 25 11:20:52.205: INFO: Created: latency-svc-f48gp + Feb 25 11:20:52.231: INFO: Created: latency-svc-8t9f2 + Feb 25 11:20:52.259: INFO: Got endpoints: latency-svc-n768j [690.875275ms] + Feb 25 11:20:52.301: INFO: Created: latency-svc-jcg9s + Feb 25 11:20:52.307: INFO: Got endpoints: latency-svc-gfjjd [718.849847ms] + Feb 25 11:20:52.328: INFO: Created: latency-svc-jmxcd + Feb 25 11:20:52.351: INFO: Got endpoints: latency-svc-bq2vf [723.403632ms] + Feb 25 11:20:52.374: INFO: Created: latency-svc-96wqv + Feb 25 11:20:52.402: INFO: Got endpoints: latency-svc-v8k24 [730.412273ms] + Feb 25 11:20:52.422: INFO: Created: latency-svc-sqtdh + Feb 25 11:20:52.450: INFO: Got endpoints: latency-svc-7rc9d [723.312363ms] + Feb 25 11:20:52.469: INFO: Created: latency-svc-fbl46 + Feb 25 11:20:52.498: INFO: Got endpoints: latency-svc-njwkm [738.011572ms] + Feb 25 11:20:52.522: INFO: Created: latency-svc-c9cmc + Feb 25 11:20:52.558: INFO: Got endpoints: latency-svc-62jxp [748.501112ms] + Feb 25 11:20:52.574: INFO: Created: latency-svc-mj5m6 + Feb 25 11:20:52.599: INFO: Got endpoints: latency-svc-zdk9d [737.057642ms] + Feb 25 11:20:52.620: INFO: Created: latency-svc-kpndk + Feb 25 11:20:52.655: INFO: Got endpoints: latency-svc-pbsz7 [752.025709ms] + Feb 25 11:20:52.701: INFO: Got endpoints: latency-svc-82z5r [733.799554ms] + Feb 25 11:20:52.756: INFO: Got endpoints: latency-svc-fdn4m [749.445868ms] + Feb 25 11:20:52.802: INFO: Got endpoints: latency-svc-7cvnc [737.692225ms] + Feb 25 11:20:52.942: INFO: Got endpoints: latency-svc-k7r76 [829.884512ms] + Feb 25 11:20:52.943: INFO: Got endpoints: latency-svc-f48gp [784.991143ms] + Feb 25 11:20:52.970: INFO: Got endpoints: latency-svc-8t9f2 [769.406404ms] + Feb 25 11:20:53.000: INFO: Got endpoints: latency-svc-jcg9s [740.537287ms] + Feb 25 11:20:53.063: INFO: Got endpoints: latency-svc-jmxcd [755.997533ms] + Feb 25 11:20:53.124: INFO: Got endpoints: latency-svc-96wqv [772.614947ms] + Feb 25 11:20:53.157: INFO: Got endpoints: latency-svc-sqtdh [754.689275ms] + Feb 25 11:20:53.239: INFO: Got endpoints: latency-svc-fbl46 [788.353429ms] + Feb 25 11:20:53.290: INFO: Got endpoints: latency-svc-c9cmc [792.01337ms] + Feb 25 11:20:53.308: INFO: Got endpoints: latency-svc-mj5m6 [749.774431ms] + Feb 25 11:20:53.367: INFO: Got endpoints: latency-svc-kpndk [767.191474ms] + Feb 25 11:20:53.367: INFO: Latencies: [60.695301ms 91.0181ms 132.993021ms 147.591886ms 156.861217ms 188.803702ms 192.281607ms 198.289856ms 201.693337ms 202.787455ms 219.981165ms 220.531853ms 223.755658ms 230.071141ms 234.712014ms 238.224937ms 265.876689ms 266.527831ms 277.774752ms 281.373826ms 289.733594ms 300.913142ms 301.760045ms 319.513827ms 324.573294ms 330.08416ms 336.955898ms 337.680466ms 353.01572ms 361.321313ms 367.800143ms 378.509026ms 383.043606ms 383.831921ms 389.166989ms 397.483689ms 414.149226ms 414.226701ms 421.524152ms 425.75944ms 442.647971ms 443.913612ms 444.85583ms 456.128882ms 459.492615ms 462.124108ms 463.335572ms 464.133831ms 468.541479ms 468.682201ms 497.921371ms 499.948498ms 500.15417ms 500.967978ms 500.983054ms 501.970486ms 509.663951ms 510.001767ms 517.548535ms 524.179702ms 527.312106ms 528.739945ms 553.514728ms 558.160651ms 563.384289ms 571.529398ms 590.043309ms 617.964979ms 633.554587ms 638.626579ms 648.721201ms 654.40904ms 666.888755ms 670.400875ms 676.987651ms 685.868755ms 687.604145ms 690.875275ms 693.691925ms 693.825287ms 695.154883ms 698.154723ms 705.947396ms 706.630763ms 706.883466ms 708.023736ms 718.745262ms 718.849847ms 723.312363ms 723.403632ms 725.628793ms 726.724858ms 727.204527ms 728.345793ms 729.512291ms 730.412273ms 730.766652ms 731.494612ms 733.799554ms 734.353982ms 735.160008ms 736.928668ms 737.057642ms 737.692225ms 738.011572ms 738.460496ms 738.488577ms 738.6959ms 740.537287ms 740.59616ms 741.092133ms 742.089109ms 742.323759ms 742.57398ms 742.662862ms 743.392974ms 744.488755ms 745.989859ms 746.457775ms 746.709132ms 746.917144ms 747.42446ms 747.562171ms 748.104799ms 748.501112ms 748.715235ms 748.725664ms 748.780828ms 748.983041ms 749.445868ms 749.494956ms 749.695502ms 749.774431ms 751.810695ms 752.025709ms 752.251462ms 752.348553ms 752.858978ms 753.202628ms 754.689275ms 754.942955ms 755.220786ms 755.454405ms 755.853327ms 755.854578ms 755.997533ms 756.091386ms 756.433044ms 757.261434ms 757.263385ms 758.811017ms 759.472115ms 759.613512ms 759.632921ms 759.649164ms 759.843702ms 759.874548ms 760.345211ms 761.920419ms 763.628966ms 763.979765ms 765.60524ms 765.855317ms 766.141586ms 767.191474ms 767.246914ms 769.406404ms 769.650509ms 771.387184ms 771.783641ms 772.213281ms 772.529815ms 772.614947ms 776.956438ms 777.032728ms 777.318469ms 777.327352ms 779.65108ms 781.160524ms 781.312834ms 784.991143ms 787.24503ms 788.353429ms 792.01337ms 795.309235ms 796.903663ms 801.384971ms 802.209556ms 809.196576ms 810.286136ms 829.884512ms 854.284573ms 854.996255ms 862.561566ms 873.963924ms 902.843147ms 905.540065ms 915.840616ms 928.629602ms 947.858421ms] + Feb 25 11:20:53.367: INFO: 50 %ile: 735.160008ms + Feb 25 11:20:53.367: INFO: 90 %ile: 784.991143ms + Feb 25 11:20:53.367: INFO: 99 %ile: 928.629602ms + Feb 25 11:20:53.367: INFO: Total sample count: 200 + [AfterEach] [sig-network] Service endpoints latency test/e2e/framework/framework.go:187 - Dec 14 16:53:58.238: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "configmap-8437" for this suite. 12/14/22 16:53:58.243 + Feb 25 11:20:53.367: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "svc-latency-4048" for this suite. 02/25/23 11:20:53.379 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] ResourceQuota - should create a ResourceQuota and capture the life of a pod. [Conformance] - test/e2e/apimachinery/resource_quota.go:220 -[BeforeEach] [sig-api-machinery] ResourceQuota +[sig-network] DNS + should provide DNS for services [Conformance] + test/e2e/network/dns.go:137 +[BeforeEach] [sig-network] DNS test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:53:58.26 -Dec 14 16:53:58.261: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename resourcequota 12/14/22 16:53:58.263 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:58.302 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:58.305 -[It] should create a ResourceQuota and capture the life of a pod. [Conformance] - test/e2e/apimachinery/resource_quota.go:220 -STEP: Counting existing ResourceQuota 12/14/22 16:53:58.31 -STEP: Creating a ResourceQuota 12/14/22 16:54:03.317 -STEP: Ensuring resource quota status is calculated 12/14/22 16:54:03.326 -STEP: Creating a Pod that fits quota 12/14/22 16:54:05.332 -STEP: Ensuring ResourceQuota status captures the pod usage 12/14/22 16:54:05.353 -STEP: Not allowing a pod to be created that exceeds remaining quota 12/14/22 16:54:07.364 -STEP: Not allowing a pod to be created that exceeds remaining quota(validation on extended resources) 12/14/22 16:54:07.369 -STEP: Ensuring a pod cannot update its resource requirements 12/14/22 16:54:07.372 -STEP: Ensuring attempts to update pod resource requirements did not change quota usage 12/14/22 16:54:07.38 -STEP: Deleting the pod 12/14/22 16:54:09.387 -STEP: Ensuring resource quota status released the pod usage 12/14/22 16:54:09.408 -[AfterEach] [sig-api-machinery] ResourceQuota +STEP: Creating a kubernetes client 02/25/23 11:20:53.398 +Feb 25 11:20:53.398: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename dns 02/25/23 11:20:53.401 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:53.449 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:53.456 +[It] should provide DNS for services [Conformance] + test/e2e/network/dns.go:137 +STEP: Creating a test headless service 02/25/23 11:20:53.459 +STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-3575.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-3575.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-3575.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-3575.svc.cluster.local;check="$$(dig +notcp +noall +answer +search 151.35.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.35.151_udp@PTR;check="$$(dig +tcp +noall +answer +search 151.35.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.35.151_tcp@PTR;sleep 1; done + 02/25/23 11:20:53.51 +STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-3575.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-3575.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-3575.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-3575.svc.cluster.local;check="$$(dig +notcp +noall +answer +search 151.35.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.35.151_udp@PTR;check="$$(dig +tcp +noall +answer +search 151.35.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.35.151_tcp@PTR;sleep 1; done + 02/25/23 11:20:53.51 +STEP: creating a pod to probe DNS 02/25/23 11:20:53.51 +STEP: submitting the pod to kubernetes 02/25/23 11:20:53.511 +Feb 25 11:20:53.569: INFO: Waiting up to 15m0s for pod "dns-test-f849f207-71c9-4ced-8387-6a68d98fc726" in namespace "dns-3575" to be "running" +Feb 25 11:20:53.604: INFO: Pod "dns-test-f849f207-71c9-4ced-8387-6a68d98fc726": Phase="Pending", Reason="", readiness=false. Elapsed: 35.028916ms +Feb 25 11:20:55.615: INFO: Pod "dns-test-f849f207-71c9-4ced-8387-6a68d98fc726": Phase="Pending", Reason="", readiness=false. Elapsed: 2.04601358s +Feb 25 11:20:57.613: INFO: Pod "dns-test-f849f207-71c9-4ced-8387-6a68d98fc726": Phase="Running", Reason="", readiness=true. Elapsed: 4.044249332s +Feb 25 11:20:57.613: INFO: Pod "dns-test-f849f207-71c9-4ced-8387-6a68d98fc726" satisfied condition "running" +STEP: retrieving the pod 02/25/23 11:20:57.613 +STEP: looking for the results for each expected name from probers 02/25/23 11:20:57.621 +Feb 25 11:20:57.631: INFO: Unable to read wheezy_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:20:57.637: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:20:57.643: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:20:57.652: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:20:57.687: INFO: Unable to read jessie_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:20:57.692: INFO: Unable to read jessie_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:20:57.697: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:20:57.703: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:20:57.726: INFO: Lookups using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 failed for: [wheezy_udp@dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_udp@dns-test-service.dns-3575.svc.cluster.local jessie_tcp@dns-test-service.dns-3575.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local] + +Feb 25 11:21:02.736: INFO: Unable to read wheezy_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:02.765: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:02.777: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:02.795: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:02.848: INFO: Unable to read jessie_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:02.863: INFO: Unable to read jessie_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:02.878: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:02.930: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:03.012: INFO: Lookups using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 failed for: [wheezy_udp@dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_udp@dns-test-service.dns-3575.svc.cluster.local jessie_tcp@dns-test-service.dns-3575.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local] + +Feb 25 11:21:07.741: INFO: Unable to read wheezy_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:07.754: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:07.773: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:07.787: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:07.871: INFO: Unable to read jessie_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:07.882: INFO: Unable to read jessie_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:07.888: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:07.895: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:08.091: INFO: Lookups using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 failed for: [wheezy_udp@dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_udp@dns-test-service.dns-3575.svc.cluster.local jessie_tcp@dns-test-service.dns-3575.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local] + +Feb 25 11:21:12.740: INFO: Unable to read wheezy_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:12.747: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:12.752: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:12.758: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:12.793: INFO: Unable to read jessie_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:12.800: INFO: Unable to read jessie_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:12.814: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:12.818: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:12.855: INFO: Lookups using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 failed for: [wheezy_udp@dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_udp@dns-test-service.dns-3575.svc.cluster.local jessie_tcp@dns-test-service.dns-3575.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local] + +Feb 25 11:21:17.733: INFO: Unable to read wheezy_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:17.739: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:17.745: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:17.749: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:17.776: INFO: Unable to read jessie_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:17.781: INFO: Unable to read jessie_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:17.787: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:17.792: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:17.816: INFO: Lookups using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 failed for: [wheezy_udp@dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_udp@dns-test-service.dns-3575.svc.cluster.local jessie_tcp@dns-test-service.dns-3575.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local] + +Feb 25 11:21:22.733: INFO: Unable to read wheezy_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:22.742: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:22.753: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:22.764: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:22.803: INFO: Unable to read jessie_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:22.809: INFO: Unable to read jessie_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:22.814: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:22.820: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:22.846: INFO: Lookups using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 failed for: [wheezy_udp@dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_udp@dns-test-service.dns-3575.svc.cluster.local jessie_tcp@dns-test-service.dns-3575.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local] + +Feb 25 11:21:27.758: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) +Feb 25 11:21:27.850: INFO: Lookups using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 failed for: [wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local] + +Feb 25 11:21:32.979: INFO: DNS probes using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 succeeded + +STEP: deleting the pod 02/25/23 11:21:32.979 +STEP: deleting the test service 02/25/23 11:21:33.01 +STEP: deleting the test headless service 02/25/23 11:21:33.321 +[AfterEach] [sig-network] DNS test/e2e/framework/framework.go:187 -Dec 14 16:54:11.417: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-8878" for this suite. 12/14/22 16:54:11.423 -{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a pod. [Conformance]","completed":276,"skipped":5072,"failed":0} +Feb 25 11:21:33.423: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "dns-3575" for this suite. 02/25/23 11:21:33.455 +{"msg":"PASSED [sig-network] DNS should provide DNS for services [Conformance]","completed":273,"skipped":5221,"failed":0} ------------------------------ -• [SLOW TEST] [13.172 seconds] -[sig-api-machinery] ResourceQuota -test/e2e/apimachinery/framework.go:23 - should create a ResourceQuota and capture the life of a pod. [Conformance] - test/e2e/apimachinery/resource_quota.go:220 +• [SLOW TEST] [40.078 seconds] +[sig-network] DNS +test/e2e/network/common/framework.go:23 + should provide DNS for services [Conformance] + test/e2e/network/dns.go:137 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] ResourceQuota + [BeforeEach] [sig-network] DNS test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:53:58.26 - Dec 14 16:53:58.261: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename resourcequota 12/14/22 16:53:58.263 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:53:58.302 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:53:58.305 - [It] should create a ResourceQuota and capture the life of a pod. [Conformance] - test/e2e/apimachinery/resource_quota.go:220 - STEP: Counting existing ResourceQuota 12/14/22 16:53:58.31 - STEP: Creating a ResourceQuota 12/14/22 16:54:03.317 - STEP: Ensuring resource quota status is calculated 12/14/22 16:54:03.326 - STEP: Creating a Pod that fits quota 12/14/22 16:54:05.332 - STEP: Ensuring ResourceQuota status captures the pod usage 12/14/22 16:54:05.353 - STEP: Not allowing a pod to be created that exceeds remaining quota 12/14/22 16:54:07.364 - STEP: Not allowing a pod to be created that exceeds remaining quota(validation on extended resources) 12/14/22 16:54:07.369 - STEP: Ensuring a pod cannot update its resource requirements 12/14/22 16:54:07.372 - STEP: Ensuring attempts to update pod resource requirements did not change quota usage 12/14/22 16:54:07.38 - STEP: Deleting the pod 12/14/22 16:54:09.387 - STEP: Ensuring resource quota status released the pod usage 12/14/22 16:54:09.408 - [AfterEach] [sig-api-machinery] ResourceQuota + STEP: Creating a kubernetes client 02/25/23 11:20:53.398 + Feb 25 11:20:53.398: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename dns 02/25/23 11:20:53.401 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:20:53.449 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:20:53.456 + [It] should provide DNS for services [Conformance] + test/e2e/network/dns.go:137 + STEP: Creating a test headless service 02/25/23 11:20:53.459 + STEP: Running these commands on wheezy: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-3575.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_udp@dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-3575.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_udp@_http._tcp.test-service-2.dns-3575.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/wheezy_tcp@_http._tcp.test-service-2.dns-3575.svc.cluster.local;check="$$(dig +notcp +noall +answer +search 151.35.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.35.151_udp@PTR;check="$$(dig +tcp +noall +answer +search 151.35.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.35.151_tcp@PTR;sleep 1; done + 02/25/23 11:20:53.51 + STEP: Running these commands on jessie: for i in `seq 1 600`; do check="$$(dig +notcp +noall +answer +search dns-test-service.dns-3575.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_udp@dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +tcp +noall +answer +search dns-test-service.dns-3575.svc.cluster.local A)" && test -n "$$check" && echo OK > /results/jessie_tcp@dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.dns-test-service.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.dns-test-service.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local;check="$$(dig +notcp +noall +answer +search _http._tcp.test-service-2.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_udp@_http._tcp.test-service-2.dns-3575.svc.cluster.local;check="$$(dig +tcp +noall +answer +search _http._tcp.test-service-2.dns-3575.svc.cluster.local SRV)" && test -n "$$check" && echo OK > /results/jessie_tcp@_http._tcp.test-service-2.dns-3575.svc.cluster.local;check="$$(dig +notcp +noall +answer +search 151.35.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.35.151_udp@PTR;check="$$(dig +tcp +noall +answer +search 151.35.233.10.in-addr.arpa. PTR)" && test -n "$$check" && echo OK > /results/10.233.35.151_tcp@PTR;sleep 1; done + 02/25/23 11:20:53.51 + STEP: creating a pod to probe DNS 02/25/23 11:20:53.51 + STEP: submitting the pod to kubernetes 02/25/23 11:20:53.511 + Feb 25 11:20:53.569: INFO: Waiting up to 15m0s for pod "dns-test-f849f207-71c9-4ced-8387-6a68d98fc726" in namespace "dns-3575" to be "running" + Feb 25 11:20:53.604: INFO: Pod "dns-test-f849f207-71c9-4ced-8387-6a68d98fc726": Phase="Pending", Reason="", readiness=false. Elapsed: 35.028916ms + Feb 25 11:20:55.615: INFO: Pod "dns-test-f849f207-71c9-4ced-8387-6a68d98fc726": Phase="Pending", Reason="", readiness=false. Elapsed: 2.04601358s + Feb 25 11:20:57.613: INFO: Pod "dns-test-f849f207-71c9-4ced-8387-6a68d98fc726": Phase="Running", Reason="", readiness=true. Elapsed: 4.044249332s + Feb 25 11:20:57.613: INFO: Pod "dns-test-f849f207-71c9-4ced-8387-6a68d98fc726" satisfied condition "running" + STEP: retrieving the pod 02/25/23 11:20:57.613 + STEP: looking for the results for each expected name from probers 02/25/23 11:20:57.621 + Feb 25 11:20:57.631: INFO: Unable to read wheezy_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:20:57.637: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:20:57.643: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:20:57.652: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:20:57.687: INFO: Unable to read jessie_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:20:57.692: INFO: Unable to read jessie_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:20:57.697: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:20:57.703: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:20:57.726: INFO: Lookups using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 failed for: [wheezy_udp@dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_udp@dns-test-service.dns-3575.svc.cluster.local jessie_tcp@dns-test-service.dns-3575.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local] + + Feb 25 11:21:02.736: INFO: Unable to read wheezy_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:02.765: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:02.777: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:02.795: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:02.848: INFO: Unable to read jessie_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:02.863: INFO: Unable to read jessie_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:02.878: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:02.930: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:03.012: INFO: Lookups using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 failed for: [wheezy_udp@dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_udp@dns-test-service.dns-3575.svc.cluster.local jessie_tcp@dns-test-service.dns-3575.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local] + + Feb 25 11:21:07.741: INFO: Unable to read wheezy_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:07.754: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:07.773: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:07.787: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:07.871: INFO: Unable to read jessie_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:07.882: INFO: Unable to read jessie_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:07.888: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:07.895: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:08.091: INFO: Lookups using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 failed for: [wheezy_udp@dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_udp@dns-test-service.dns-3575.svc.cluster.local jessie_tcp@dns-test-service.dns-3575.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local] + + Feb 25 11:21:12.740: INFO: Unable to read wheezy_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:12.747: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:12.752: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:12.758: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:12.793: INFO: Unable to read jessie_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:12.800: INFO: Unable to read jessie_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:12.814: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:12.818: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:12.855: INFO: Lookups using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 failed for: [wheezy_udp@dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_udp@dns-test-service.dns-3575.svc.cluster.local jessie_tcp@dns-test-service.dns-3575.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local] + + Feb 25 11:21:17.733: INFO: Unable to read wheezy_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:17.739: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:17.745: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:17.749: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:17.776: INFO: Unable to read jessie_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:17.781: INFO: Unable to read jessie_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:17.787: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:17.792: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:17.816: INFO: Lookups using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 failed for: [wheezy_udp@dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_udp@dns-test-service.dns-3575.svc.cluster.local jessie_tcp@dns-test-service.dns-3575.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local] + + Feb 25 11:21:22.733: INFO: Unable to read wheezy_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:22.742: INFO: Unable to read wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:22.753: INFO: Unable to read wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:22.764: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:22.803: INFO: Unable to read jessie_udp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:22.809: INFO: Unable to read jessie_tcp@dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:22.814: INFO: Unable to read jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:22.820: INFO: Unable to read jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:22.846: INFO: Lookups using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 failed for: [wheezy_udp@dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@dns-test-service.dns-3575.svc.cluster.local wheezy_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_udp@dns-test-service.dns-3575.svc.cluster.local jessie_tcp@dns-test-service.dns-3575.svc.cluster.local jessie_udp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local jessie_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local] + + Feb 25 11:21:27.758: INFO: Unable to read wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local from pod dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726: the server could not find the requested resource (get pods dns-test-f849f207-71c9-4ced-8387-6a68d98fc726) + Feb 25 11:21:27.850: INFO: Lookups using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 failed for: [wheezy_tcp@_http._tcp.dns-test-service.dns-3575.svc.cluster.local] + + Feb 25 11:21:32.979: INFO: DNS probes using dns-3575/dns-test-f849f207-71c9-4ced-8387-6a68d98fc726 succeeded + + STEP: deleting the pod 02/25/23 11:21:32.979 + STEP: deleting the test service 02/25/23 11:21:33.01 + STEP: deleting the test headless service 02/25/23 11:21:33.321 + [AfterEach] [sig-network] DNS test/e2e/framework/framework.go:187 - Dec 14 16:54:11.417: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "resourcequota-8878" for this suite. 12/14/22 16:54:11.423 + Feb 25 11:21:33.423: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "dns-3575" for this suite. 02/25/23 11:21:33.455 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSS ------------------------------ -[sig-apps] ReplicaSet - should adopt matching pods on creation and release no longer matching pods [Conformance] - test/e2e/apps/replica_set.go:131 -[BeforeEach] [sig-apps] ReplicaSet +[sig-storage] Secrets + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:78 +[BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:54:11.464 -Dec 14 16:54:11.464: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename replicaset 12/14/22 16:54:11.465 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:54:11.493 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:54:11.496 -[It] should adopt matching pods on creation and release no longer matching pods [Conformance] - test/e2e/apps/replica_set.go:131 -STEP: Given a Pod with a 'name' label pod-adoption-release is created 12/14/22 16:54:11.499 -Dec 14 16:54:11.523: INFO: Waiting up to 5m0s for pod "pod-adoption-release" in namespace "replicaset-4553" to be "running and ready" -Dec 14 16:54:11.529: INFO: Pod "pod-adoption-release": Phase="Pending", Reason="", readiness=false. Elapsed: 6.412594ms -Dec 14 16:54:11.530: INFO: The phase of Pod pod-adoption-release is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:54:13.540: INFO: Pod "pod-adoption-release": Phase="Running", Reason="", readiness=true. Elapsed: 2.017153403s -Dec 14 16:54:13.540: INFO: The phase of Pod pod-adoption-release is Running (Ready = true) -Dec 14 16:54:13.540: INFO: Pod "pod-adoption-release" satisfied condition "running and ready" -STEP: When a replicaset with a matching selector is created 12/14/22 16:54:13.545 -STEP: Then the orphan pod is adopted 12/14/22 16:54:13.553 -STEP: When the matched label of one of its pods change 12/14/22 16:54:14.568 -Dec 14 16:54:14.573: INFO: Pod name pod-adoption-release: Found 1 pods out of 1 -STEP: Then the pod is released 12/14/22 16:54:14.59 -[AfterEach] [sig-apps] ReplicaSet +STEP: Creating a kubernetes client 02/25/23 11:21:33.491 +Feb 25 11:21:33.491: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename secrets 02/25/23 11:21:33.495 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:21:33.533 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:21:33.542 +[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:78 +STEP: Creating secret with name secret-test-map-a70097cc-ce4a-4b63-a2e8-7e3278579e14 02/25/23 11:21:33.55 +STEP: Creating a pod to test consume secrets 02/25/23 11:21:33.561 +Feb 25 11:21:33.580: INFO: Waiting up to 5m0s for pod "pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c" in namespace "secrets-7920" to be "Succeeded or Failed" +Feb 25 11:21:33.588: INFO: Pod "pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c": Phase="Pending", Reason="", readiness=false. Elapsed: 7.82965ms +Feb 25 11:21:35.599: INFO: Pod "pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018650271s +Feb 25 11:21:37.597: INFO: Pod "pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016443383s +STEP: Saw pod success 02/25/23 11:21:37.597 +Feb 25 11:21:37.598: INFO: Pod "pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c" satisfied condition "Succeeded or Failed" +Feb 25 11:21:37.603: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c container secret-volume-test: +STEP: delete the pod 02/25/23 11:21:37.617 +Feb 25 11:21:37.638: INFO: Waiting for pod pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c to disappear +Feb 25 11:21:37.644: INFO: Pod pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c no longer exists +[AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 -Dec 14 16:54:15.602: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "replicaset-4553" for this suite. 12/14/22 16:54:15.609 -{"msg":"PASSED [sig-apps] ReplicaSet should adopt matching pods on creation and release no longer matching pods [Conformance]","completed":277,"skipped":5258,"failed":0} +Feb 25 11:21:37.644: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-7920" for this suite. 02/25/23 11:21:37.653 +{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","completed":274,"skipped":5237,"failed":0} ------------------------------ -• [4.153 seconds] -[sig-apps] ReplicaSet -test/e2e/apps/framework.go:23 - should adopt matching pods on creation and release no longer matching pods [Conformance] - test/e2e/apps/replica_set.go:131 +• [4.176 seconds] +[sig-storage] Secrets +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:78 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] ReplicaSet + [BeforeEach] [sig-storage] Secrets test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:54:11.464 - Dec 14 16:54:11.464: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename replicaset 12/14/22 16:54:11.465 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:54:11.493 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:54:11.496 - [It] should adopt matching pods on creation and release no longer matching pods [Conformance] - test/e2e/apps/replica_set.go:131 - STEP: Given a Pod with a 'name' label pod-adoption-release is created 12/14/22 16:54:11.499 - Dec 14 16:54:11.523: INFO: Waiting up to 5m0s for pod "pod-adoption-release" in namespace "replicaset-4553" to be "running and ready" - Dec 14 16:54:11.529: INFO: Pod "pod-adoption-release": Phase="Pending", Reason="", readiness=false. Elapsed: 6.412594ms - Dec 14 16:54:11.530: INFO: The phase of Pod pod-adoption-release is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:54:13.540: INFO: Pod "pod-adoption-release": Phase="Running", Reason="", readiness=true. Elapsed: 2.017153403s - Dec 14 16:54:13.540: INFO: The phase of Pod pod-adoption-release is Running (Ready = true) - Dec 14 16:54:13.540: INFO: Pod "pod-adoption-release" satisfied condition "running and ready" - STEP: When a replicaset with a matching selector is created 12/14/22 16:54:13.545 - STEP: Then the orphan pod is adopted 12/14/22 16:54:13.553 - STEP: When the matched label of one of its pods change 12/14/22 16:54:14.568 - Dec 14 16:54:14.573: INFO: Pod name pod-adoption-release: Found 1 pods out of 1 - STEP: Then the pod is released 12/14/22 16:54:14.59 - [AfterEach] [sig-apps] ReplicaSet + STEP: Creating a kubernetes client 02/25/23 11:21:33.491 + Feb 25 11:21:33.491: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename secrets 02/25/23 11:21:33.495 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:21:33.533 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:21:33.542 + [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/secrets_volume.go:78 + STEP: Creating secret with name secret-test-map-a70097cc-ce4a-4b63-a2e8-7e3278579e14 02/25/23 11:21:33.55 + STEP: Creating a pod to test consume secrets 02/25/23 11:21:33.561 + Feb 25 11:21:33.580: INFO: Waiting up to 5m0s for pod "pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c" in namespace "secrets-7920" to be "Succeeded or Failed" + Feb 25 11:21:33.588: INFO: Pod "pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c": Phase="Pending", Reason="", readiness=false. Elapsed: 7.82965ms + Feb 25 11:21:35.599: INFO: Pod "pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018650271s + Feb 25 11:21:37.597: INFO: Pod "pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.016443383s + STEP: Saw pod success 02/25/23 11:21:37.597 + Feb 25 11:21:37.598: INFO: Pod "pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c" satisfied condition "Succeeded or Failed" + Feb 25 11:21:37.603: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c container secret-volume-test: + STEP: delete the pod 02/25/23 11:21:37.617 + Feb 25 11:21:37.638: INFO: Waiting for pod pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c to disappear + Feb 25 11:21:37.644: INFO: Pod pod-secrets-4f8cd3b1-35e2-4e96-ac34-7628b88d738c no longer exists + [AfterEach] [sig-storage] Secrets test/e2e/framework/framework.go:187 - Dec 14 16:54:15.602: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "replicaset-4553" for this suite. 12/14/22 16:54:15.609 + Feb 25 11:21:37.644: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "secrets-7920" for this suite. 02/25/23 11:21:37.653 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Containers - should use the image defaults if command and args are blank [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:38 -[BeforeEach] [sig-node] Containers +[sig-node] Variable Expansion + should allow composing env vars into new env vars [NodeConformance] [Conformance] + test/e2e/common/node/expansion.go:43 +[BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:54:15.622 -Dec 14 16:54:15.622: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename containers 12/14/22 16:54:15.624 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:54:15.651 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:54:15.657 -[It] should use the image defaults if command and args are blank [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:38 -Dec 14 16:54:15.680: INFO: Waiting up to 5m0s for pod "client-containers-d4f0c719-c9c0-42f9-b7b3-17aed4eef313" in namespace "containers-2824" to be "running" -Dec 14 16:54:15.689: INFO: Pod "client-containers-d4f0c719-c9c0-42f9-b7b3-17aed4eef313": Phase="Pending", Reason="", readiness=false. Elapsed: 8.52453ms -Dec 14 16:54:17.697: INFO: Pod "client-containers-d4f0c719-c9c0-42f9-b7b3-17aed4eef313": Phase="Running", Reason="", readiness=true. Elapsed: 2.016824772s -Dec 14 16:54:17.697: INFO: Pod "client-containers-d4f0c719-c9c0-42f9-b7b3-17aed4eef313" satisfied condition "running" -[AfterEach] [sig-node] Containers +STEP: Creating a kubernetes client 02/25/23 11:21:37.673 +Feb 25 11:21:37.673: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename var-expansion 02/25/23 11:21:37.675 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:21:37.698 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:21:37.703 +[It] should allow composing env vars into new env vars [NodeConformance] [Conformance] + test/e2e/common/node/expansion.go:43 +STEP: Creating a pod to test env composition 02/25/23 11:21:37.707 +Feb 25 11:21:37.718: INFO: Waiting up to 5m0s for pod "var-expansion-9b630f7b-8146-4257-9007-91a350f4d987" in namespace "var-expansion-5389" to be "Succeeded or Failed" +Feb 25 11:21:37.728: INFO: Pod "var-expansion-9b630f7b-8146-4257-9007-91a350f4d987": Phase="Pending", Reason="", readiness=false. Elapsed: 9.316581ms +Feb 25 11:21:39.736: INFO: Pod "var-expansion-9b630f7b-8146-4257-9007-91a350f4d987": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017752377s +Feb 25 11:21:41.736: INFO: Pod "var-expansion-9b630f7b-8146-4257-9007-91a350f4d987": Phase="Pending", Reason="", readiness=false. Elapsed: 4.017757226s +Feb 25 11:21:43.735: INFO: Pod "var-expansion-9b630f7b-8146-4257-9007-91a350f4d987": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016292251s +STEP: Saw pod success 02/25/23 11:21:43.735 +Feb 25 11:21:43.735: INFO: Pod "var-expansion-9b630f7b-8146-4257-9007-91a350f4d987" satisfied condition "Succeeded or Failed" +Feb 25 11:21:43.739: INFO: Trying to get logs from node diek3dua9gay-3 pod var-expansion-9b630f7b-8146-4257-9007-91a350f4d987 container dapi-container: +STEP: delete the pod 02/25/23 11:21:44.078 +Feb 25 11:21:44.096: INFO: Waiting for pod var-expansion-9b630f7b-8146-4257-9007-91a350f4d987 to disappear +Feb 25 11:21:44.103: INFO: Pod var-expansion-9b630f7b-8146-4257-9007-91a350f4d987 no longer exists +[AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 -Dec 14 16:54:17.705: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "containers-2824" for this suite. 12/14/22 16:54:17.712 -{"msg":"PASSED [sig-node] Containers should use the image defaults if command and args are blank [NodeConformance] [Conformance]","completed":278,"skipped":5274,"failed":0} +Feb 25 11:21:44.103: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "var-expansion-5389" for this suite. 02/25/23 11:21:44.111 +{"msg":"PASSED [sig-node] Variable Expansion should allow composing env vars into new env vars [NodeConformance] [Conformance]","completed":275,"skipped":5254,"failed":0} ------------------------------ -• [2.098 seconds] -[sig-node] Containers +• [SLOW TEST] [6.451 seconds] +[sig-node] Variable Expansion test/e2e/common/node/framework.go:23 - should use the image defaults if command and args are blank [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:38 + should allow composing env vars into new env vars [NodeConformance] [Conformance] + test/e2e/common/node/expansion.go:43 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Containers + [BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:54:15.622 - Dec 14 16:54:15.622: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename containers 12/14/22 16:54:15.624 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:54:15.651 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:54:15.657 - [It] should use the image defaults if command and args are blank [NodeConformance] [Conformance] - test/e2e/common/node/containers.go:38 - Dec 14 16:54:15.680: INFO: Waiting up to 5m0s for pod "client-containers-d4f0c719-c9c0-42f9-b7b3-17aed4eef313" in namespace "containers-2824" to be "running" - Dec 14 16:54:15.689: INFO: Pod "client-containers-d4f0c719-c9c0-42f9-b7b3-17aed4eef313": Phase="Pending", Reason="", readiness=false. Elapsed: 8.52453ms - Dec 14 16:54:17.697: INFO: Pod "client-containers-d4f0c719-c9c0-42f9-b7b3-17aed4eef313": Phase="Running", Reason="", readiness=true. Elapsed: 2.016824772s - Dec 14 16:54:17.697: INFO: Pod "client-containers-d4f0c719-c9c0-42f9-b7b3-17aed4eef313" satisfied condition "running" - [AfterEach] [sig-node] Containers + STEP: Creating a kubernetes client 02/25/23 11:21:37.673 + Feb 25 11:21:37.673: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename var-expansion 02/25/23 11:21:37.675 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:21:37.698 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:21:37.703 + [It] should allow composing env vars into new env vars [NodeConformance] [Conformance] + test/e2e/common/node/expansion.go:43 + STEP: Creating a pod to test env composition 02/25/23 11:21:37.707 + Feb 25 11:21:37.718: INFO: Waiting up to 5m0s for pod "var-expansion-9b630f7b-8146-4257-9007-91a350f4d987" in namespace "var-expansion-5389" to be "Succeeded or Failed" + Feb 25 11:21:37.728: INFO: Pod "var-expansion-9b630f7b-8146-4257-9007-91a350f4d987": Phase="Pending", Reason="", readiness=false. Elapsed: 9.316581ms + Feb 25 11:21:39.736: INFO: Pod "var-expansion-9b630f7b-8146-4257-9007-91a350f4d987": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017752377s + Feb 25 11:21:41.736: INFO: Pod "var-expansion-9b630f7b-8146-4257-9007-91a350f4d987": Phase="Pending", Reason="", readiness=false. Elapsed: 4.017757226s + Feb 25 11:21:43.735: INFO: Pod "var-expansion-9b630f7b-8146-4257-9007-91a350f4d987": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.016292251s + STEP: Saw pod success 02/25/23 11:21:43.735 + Feb 25 11:21:43.735: INFO: Pod "var-expansion-9b630f7b-8146-4257-9007-91a350f4d987" satisfied condition "Succeeded or Failed" + Feb 25 11:21:43.739: INFO: Trying to get logs from node diek3dua9gay-3 pod var-expansion-9b630f7b-8146-4257-9007-91a350f4d987 container dapi-container: + STEP: delete the pod 02/25/23 11:21:44.078 + Feb 25 11:21:44.096: INFO: Waiting for pod var-expansion-9b630f7b-8146-4257-9007-91a350f4d987 to disappear + Feb 25 11:21:44.103: INFO: Pod var-expansion-9b630f7b-8146-4257-9007-91a350f4d987 no longer exists + [AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 - Dec 14 16:54:17.705: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "containers-2824" for this suite. 12/14/22 16:54:17.712 + Feb 25 11:21:44.103: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "var-expansion-5389" for this suite. 02/25/23 11:21:44.111 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSS ------------------------------ -[sig-node] Downward API - should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:165 -[BeforeEach] [sig-node] Downward API +[sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath + runs ReplicaSets to verify preemption running path [Conformance] + test/e2e/scheduling/preemption.go:543 +[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:54:17.728 -Dec 14 16:54:17.729: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename downward-api 12/14/22 16:54:17.731 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:54:17.763 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:54:17.787 -[It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:165 -STEP: Creating a pod to test downward api env vars 12/14/22 16:54:17.808 -Dec 14 16:54:17.834: INFO: Waiting up to 5m0s for pod "downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee" in namespace "downward-api-2669" to be "Succeeded or Failed" -Dec 14 16:54:17.850: INFO: Pod "downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee": Phase="Pending", Reason="", readiness=false. Elapsed: 15.397159ms -Dec 14 16:54:19.861: INFO: Pod "downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026982436s -Dec 14 16:54:21.860: INFO: Pod "downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025645856s -STEP: Saw pod success 12/14/22 16:54:21.86 -Dec 14 16:54:21.860: INFO: Pod "downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee" satisfied condition "Succeeded or Failed" -Dec 14 16:54:21.869: INFO: Trying to get logs from node kiedai4iakae-3 pod downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee container dapi-container: -STEP: delete the pod 12/14/22 16:54:22.063 -Dec 14 16:54:22.083: INFO: Waiting for pod downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee to disappear -Dec 14 16:54:22.087: INFO: Pod downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee no longer exists -[AfterEach] [sig-node] Downward API +STEP: Creating a kubernetes client 02/25/23 11:21:44.136 +Feb 25 11:21:44.136: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename sched-preemption 02/25/23 11:21:44.14 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:21:44.162 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:21:44.166 +[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:92 +Feb 25 11:21:44.191: INFO: Waiting up to 1m0s for all nodes to be ready +Feb 25 11:22:44.250: INFO: Waiting for terminating namespaces to be deleted... +[BeforeEach] PreemptionExecutionPath + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 11:22:44.256 +Feb 25 11:22:44.257: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename sched-preemption-path 02/25/23 11:22:44.266 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:22:44.285 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:22:44.289 +[BeforeEach] PreemptionExecutionPath + test/e2e/scheduling/preemption.go:496 +STEP: Finding an available node 02/25/23 11:22:44.295 +STEP: Trying to launch a pod without a label to get a node which can launch it. 02/25/23 11:22:44.295 +Feb 25 11:22:44.309: INFO: Waiting up to 1m0s for pod "without-label" in namespace "sched-preemption-path-5898" to be "running" +Feb 25 11:22:44.316: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 7.612621ms +Feb 25 11:22:46.322: INFO: Pod "without-label": Phase="Running", Reason="", readiness=true. Elapsed: 2.013566793s +Feb 25 11:22:46.323: INFO: Pod "without-label" satisfied condition "running" +STEP: Explicitly delete pod here to free the resource it takes. 02/25/23 11:22:46.328 +Feb 25 11:22:46.350: INFO: found a healthy node: diek3dua9gay-3 +[It] runs ReplicaSets to verify preemption running path [Conformance] + test/e2e/scheduling/preemption.go:543 +Feb 25 11:22:54.492: INFO: pods created so far: [1 1 1] +Feb 25 11:22:54.493: INFO: length of pods created so far: 3 +Feb 25 11:22:56.507: INFO: pods created so far: [2 2 1] +[AfterEach] PreemptionExecutionPath + test/e2e/framework/framework.go:187 +Feb 25 11:23:03.509: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "sched-preemption-path-5898" for this suite. 02/25/23 11:23:03.517 +[AfterEach] PreemptionExecutionPath + test/e2e/scheduling/preemption.go:470 +[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:187 -Dec 14 16:54:22.087: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "downward-api-2669" for this suite. 12/14/22 16:54:22.093 -{"msg":"PASSED [sig-node] Downward API should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance]","completed":279,"skipped":5295,"failed":0} +Feb 25 11:23:03.580: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "sched-preemption-9438" for this suite. 02/25/23 11:23:03.585 +[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:80 +{"msg":"PASSED [sig-scheduling] SchedulerPreemption [Serial] PreemptionExecutionPath runs ReplicaSets to verify preemption running path [Conformance]","completed":276,"skipped":5269,"failed":0} ------------------------------ -• [4.372 seconds] -[sig-node] Downward API -test/e2e/common/node/framework.go:23 - should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:165 +• [SLOW TEST] [79.548 seconds] +[sig-scheduling] SchedulerPreemption [Serial] +test/e2e/scheduling/framework.go:40 + PreemptionExecutionPath + test/e2e/scheduling/preemption.go:458 + runs ReplicaSets to verify preemption running path [Conformance] + test/e2e/scheduling/preemption.go:543 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Downward API + [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:54:17.728 - Dec 14 16:54:17.729: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename downward-api 12/14/22 16:54:17.731 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:54:17.763 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:54:17.787 - [It] should provide container's limits.cpu/memory and requests.cpu/memory as env vars [NodeConformance] [Conformance] - test/e2e/common/node/downwardapi.go:165 - STEP: Creating a pod to test downward api env vars 12/14/22 16:54:17.808 - Dec 14 16:54:17.834: INFO: Waiting up to 5m0s for pod "downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee" in namespace "downward-api-2669" to be "Succeeded or Failed" - Dec 14 16:54:17.850: INFO: Pod "downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee": Phase="Pending", Reason="", readiness=false. Elapsed: 15.397159ms - Dec 14 16:54:19.861: INFO: Pod "downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee": Phase="Pending", Reason="", readiness=false. Elapsed: 2.026982436s - Dec 14 16:54:21.860: INFO: Pod "downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.025645856s - STEP: Saw pod success 12/14/22 16:54:21.86 - Dec 14 16:54:21.860: INFO: Pod "downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee" satisfied condition "Succeeded or Failed" - Dec 14 16:54:21.869: INFO: Trying to get logs from node kiedai4iakae-3 pod downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee container dapi-container: - STEP: delete the pod 12/14/22 16:54:22.063 - Dec 14 16:54:22.083: INFO: Waiting for pod downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee to disappear - Dec 14 16:54:22.087: INFO: Pod downward-api-a1619f0f-fa54-4c42-b61d-f5c6e2c5edee no longer exists - [AfterEach] [sig-node] Downward API + STEP: Creating a kubernetes client 02/25/23 11:21:44.136 + Feb 25 11:21:44.136: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename sched-preemption 02/25/23 11:21:44.14 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:21:44.162 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:21:44.166 + [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:92 + Feb 25 11:21:44.191: INFO: Waiting up to 1m0s for all nodes to be ready + Feb 25 11:22:44.250: INFO: Waiting for terminating namespaces to be deleted... + [BeforeEach] PreemptionExecutionPath + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 11:22:44.256 + Feb 25 11:22:44.257: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename sched-preemption-path 02/25/23 11:22:44.266 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:22:44.285 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:22:44.289 + [BeforeEach] PreemptionExecutionPath + test/e2e/scheduling/preemption.go:496 + STEP: Finding an available node 02/25/23 11:22:44.295 + STEP: Trying to launch a pod without a label to get a node which can launch it. 02/25/23 11:22:44.295 + Feb 25 11:22:44.309: INFO: Waiting up to 1m0s for pod "without-label" in namespace "sched-preemption-path-5898" to be "running" + Feb 25 11:22:44.316: INFO: Pod "without-label": Phase="Pending", Reason="", readiness=false. Elapsed: 7.612621ms + Feb 25 11:22:46.322: INFO: Pod "without-label": Phase="Running", Reason="", readiness=true. Elapsed: 2.013566793s + Feb 25 11:22:46.323: INFO: Pod "without-label" satisfied condition "running" + STEP: Explicitly delete pod here to free the resource it takes. 02/25/23 11:22:46.328 + Feb 25 11:22:46.350: INFO: found a healthy node: diek3dua9gay-3 + [It] runs ReplicaSets to verify preemption running path [Conformance] + test/e2e/scheduling/preemption.go:543 + Feb 25 11:22:54.492: INFO: pods created so far: [1 1 1] + Feb 25 11:22:54.493: INFO: length of pods created so far: 3 + Feb 25 11:22:56.507: INFO: pods created so far: [2 2 1] + [AfterEach] PreemptionExecutionPath + test/e2e/framework/framework.go:187 + Feb 25 11:23:03.509: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "sched-preemption-path-5898" for this suite. 02/25/23 11:23:03.517 + [AfterEach] PreemptionExecutionPath + test/e2e/scheduling/preemption.go:470 + [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:187 - Dec 14 16:54:22.087: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "downward-api-2669" for this suite. 12/14/22 16:54:22.093 + Feb 25 11:23:03.580: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "sched-preemption-9438" for this suite. 02/25/23 11:23:03.585 + [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:80 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSS +SSSSSSSSS ------------------------------ -[sig-network] Networking Granular Checks: Pods - should function for intra-pod communication: udp [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:93 -[BeforeEach] [sig-network] Networking - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:54:22.109 -Dec 14 16:54:22.109: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename pod-network-test 12/14/22 16:54:22.11 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:54:22.137 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:54:22.14 -[It] should function for intra-pod communication: udp [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:93 -STEP: Performing setup for networking test in namespace pod-network-test-8153 12/14/22 16:54:22.143 -STEP: creating a selector 12/14/22 16:54:22.143 -STEP: Creating the service pods in kubernetes 12/14/22 16:54:22.143 -Dec 14 16:54:22.143: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable -Dec 14 16:54:22.183: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-8153" to be "running and ready" -Dec 14 16:54:22.194: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 11.845317ms -Dec 14 16:54:22.195: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:54:24.200: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 2.016986611s -Dec 14 16:54:24.200: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 16:54:26.202: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.019283075s -Dec 14 16:54:26.202: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 16:54:28.204: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.021101318s -Dec 14 16:54:28.204: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 16:54:30.201: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.018065739s -Dec 14 16:54:30.201: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 16:54:32.201: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.018556939s -Dec 14 16:54:32.201: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 16:54:34.200: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 12.017230676s -Dec 14 16:54:34.200: INFO: The phase of Pod netserver-0 is Running (Ready = true) -Dec 14 16:54:34.200: INFO: Pod "netserver-0" satisfied condition "running and ready" -Dec 14 16:54:34.204: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-8153" to be "running and ready" -Dec 14 16:54:34.208: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 4.024572ms -Dec 14 16:54:34.208: INFO: The phase of Pod netserver-1 is Running (Ready = true) -Dec 14 16:54:34.208: INFO: Pod "netserver-1" satisfied condition "running and ready" -Dec 14 16:54:34.211: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-8153" to be "running and ready" -Dec 14 16:54:34.215: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 3.360865ms -Dec 14 16:54:34.215: INFO: The phase of Pod netserver-2 is Running (Ready = true) -Dec 14 16:54:34.215: INFO: Pod "netserver-2" satisfied condition "running and ready" -STEP: Creating test pods 12/14/22 16:54:34.219 -Dec 14 16:54:34.225: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-8153" to be "running" -Dec 14 16:54:34.229: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 3.816338ms -Dec 14 16:54:36.236: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.010953887s -Dec 14 16:54:36.236: INFO: Pod "test-container-pod" satisfied condition "running" -Dec 14 16:54:36.240: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 -Dec 14 16:54:36.240: INFO: Breadth first check of 10.233.64.153 on host 192.168.121.249... -Dec 14 16:54:36.245: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.146:9080/dial?request=hostname&protocol=udp&host=10.233.64.153&port=8081&tries=1'] Namespace:pod-network-test-8153 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:54:36.245: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:54:36.246: INFO: ExecWithOptions: Clientset creation -Dec 14 16:54:36.247: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-8153/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.146%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dudp%26host%3D10.233.64.153%26port%3D8081%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) -Dec 14 16:54:36.370: INFO: Waiting for responses: map[] -Dec 14 16:54:36.371: INFO: reached 10.233.64.153 after 0/1 tries -Dec 14 16:54:36.371: INFO: Breadth first check of 10.233.65.153 on host 192.168.121.197... -Dec 14 16:54:36.377: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.146:9080/dial?request=hostname&protocol=udp&host=10.233.65.153&port=8081&tries=1'] Namespace:pod-network-test-8153 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:54:36.377: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:54:36.378: INFO: ExecWithOptions: Clientset creation -Dec 14 16:54:36.378: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-8153/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.146%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dudp%26host%3D10.233.65.153%26port%3D8081%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) -Dec 14 16:54:36.491: INFO: Waiting for responses: map[] -Dec 14 16:54:36.491: INFO: reached 10.233.65.153 after 0/1 tries -Dec 14 16:54:36.491: INFO: Breadth first check of 10.233.66.145 on host 192.168.121.244... -Dec 14 16:54:36.496: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.146:9080/dial?request=hostname&protocol=udp&host=10.233.66.145&port=8081&tries=1'] Namespace:pod-network-test-8153 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:54:36.496: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:54:36.498: INFO: ExecWithOptions: Clientset creation -Dec 14 16:54:36.498: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-8153/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.146%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dudp%26host%3D10.233.66.145%26port%3D8081%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) -Dec 14 16:54:36.645: INFO: Waiting for responses: map[] -Dec 14 16:54:36.645: INFO: reached 10.233.66.145 after 0/1 tries -Dec 14 16:54:36.645: INFO: Going to retry 0 out of 3 pods.... -[AfterEach] [sig-network] Networking +[sig-storage] Downward API volume + should provide podname only [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:52 +[BeforeEach] [sig-storage] Downward API volume + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 11:23:03.689 +Feb 25 11:23:03.689: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 11:23:03.693 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:23:03.717 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:23:03.722 +[BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 +[It] should provide podname only [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:52 +STEP: Creating a pod to test downward API volume plugin 02/25/23 11:23:03.726 +Feb 25 11:23:03.750: INFO: Waiting up to 5m0s for pod "downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128" in namespace "downward-api-8575" to be "Succeeded or Failed" +Feb 25 11:23:03.762: INFO: Pod "downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128": Phase="Pending", Reason="", readiness=false. Elapsed: 12.349531ms +Feb 25 11:23:05.770: INFO: Pod "downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019790611s +Feb 25 11:23:07.768: INFO: Pod "downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018598863s +STEP: Saw pod success 02/25/23 11:23:07.769 +Feb 25 11:23:07.769: INFO: Pod "downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128" satisfied condition "Succeeded or Failed" +Feb 25 11:23:07.774: INFO: Trying to get logs from node diek3dua9gay-2 pod downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128 container client-container: +STEP: delete the pod 02/25/23 11:23:07.795 +Feb 25 11:23:07.815: INFO: Waiting for pod downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128 to disappear +Feb 25 11:23:07.820: INFO: Pod downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128 no longer exists +[AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 -Dec 14 16:54:36.645: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pod-network-test-8153" for this suite. 12/14/22 16:54:36.653 -{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for intra-pod communication: udp [NodeConformance] [Conformance]","completed":280,"skipped":5316,"failed":0} +Feb 25 11:23:07.821: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-8575" for this suite. 02/25/23 11:23:07.827 +{"msg":"PASSED [sig-storage] Downward API volume should provide podname only [NodeConformance] [Conformance]","completed":277,"skipped":5278,"failed":0} ------------------------------ -• [SLOW TEST] [14.554 seconds] -[sig-network] Networking -test/e2e/common/network/framework.go:23 - Granular Checks: Pods - test/e2e/common/network/networking.go:32 - should function for intra-pod communication: udp [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:93 +• [4.152 seconds] +[sig-storage] Downward API volume +test/e2e/common/storage/framework.go:23 + should provide podname only [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:52 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Networking + [BeforeEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:54:22.109 - Dec 14 16:54:22.109: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename pod-network-test 12/14/22 16:54:22.11 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:54:22.137 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:54:22.14 - [It] should function for intra-pod communication: udp [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:93 - STEP: Performing setup for networking test in namespace pod-network-test-8153 12/14/22 16:54:22.143 - STEP: creating a selector 12/14/22 16:54:22.143 - STEP: Creating the service pods in kubernetes 12/14/22 16:54:22.143 - Dec 14 16:54:22.143: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable - Dec 14 16:54:22.183: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-8153" to be "running and ready" - Dec 14 16:54:22.194: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 11.845317ms - Dec 14 16:54:22.195: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:54:24.200: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 2.016986611s - Dec 14 16:54:24.200: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 16:54:26.202: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.019283075s - Dec 14 16:54:26.202: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 16:54:28.204: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.021101318s - Dec 14 16:54:28.204: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 16:54:30.201: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.018065739s - Dec 14 16:54:30.201: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 16:54:32.201: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.018556939s - Dec 14 16:54:32.201: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 16:54:34.200: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 12.017230676s - Dec 14 16:54:34.200: INFO: The phase of Pod netserver-0 is Running (Ready = true) - Dec 14 16:54:34.200: INFO: Pod "netserver-0" satisfied condition "running and ready" - Dec 14 16:54:34.204: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-8153" to be "running and ready" - Dec 14 16:54:34.208: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 4.024572ms - Dec 14 16:54:34.208: INFO: The phase of Pod netserver-1 is Running (Ready = true) - Dec 14 16:54:34.208: INFO: Pod "netserver-1" satisfied condition "running and ready" - Dec 14 16:54:34.211: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-8153" to be "running and ready" - Dec 14 16:54:34.215: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 3.360865ms - Dec 14 16:54:34.215: INFO: The phase of Pod netserver-2 is Running (Ready = true) - Dec 14 16:54:34.215: INFO: Pod "netserver-2" satisfied condition "running and ready" - STEP: Creating test pods 12/14/22 16:54:34.219 - Dec 14 16:54:34.225: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-8153" to be "running" - Dec 14 16:54:34.229: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 3.816338ms - Dec 14 16:54:36.236: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.010953887s - Dec 14 16:54:36.236: INFO: Pod "test-container-pod" satisfied condition "running" - Dec 14 16:54:36.240: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 - Dec 14 16:54:36.240: INFO: Breadth first check of 10.233.64.153 on host 192.168.121.249... - Dec 14 16:54:36.245: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.146:9080/dial?request=hostname&protocol=udp&host=10.233.64.153&port=8081&tries=1'] Namespace:pod-network-test-8153 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:54:36.245: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:54:36.246: INFO: ExecWithOptions: Clientset creation - Dec 14 16:54:36.247: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-8153/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.146%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dudp%26host%3D10.233.64.153%26port%3D8081%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) - Dec 14 16:54:36.370: INFO: Waiting for responses: map[] - Dec 14 16:54:36.371: INFO: reached 10.233.64.153 after 0/1 tries - Dec 14 16:54:36.371: INFO: Breadth first check of 10.233.65.153 on host 192.168.121.197... - Dec 14 16:54:36.377: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.146:9080/dial?request=hostname&protocol=udp&host=10.233.65.153&port=8081&tries=1'] Namespace:pod-network-test-8153 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:54:36.377: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:54:36.378: INFO: ExecWithOptions: Clientset creation - Dec 14 16:54:36.378: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-8153/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.146%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dudp%26host%3D10.233.65.153%26port%3D8081%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) - Dec 14 16:54:36.491: INFO: Waiting for responses: map[] - Dec 14 16:54:36.491: INFO: reached 10.233.65.153 after 0/1 tries - Dec 14 16:54:36.491: INFO: Breadth first check of 10.233.66.145 on host 192.168.121.244... - Dec 14 16:54:36.496: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s 'http://10.233.66.146:9080/dial?request=hostname&protocol=udp&host=10.233.66.145&port=8081&tries=1'] Namespace:pod-network-test-8153 PodName:test-container-pod ContainerName:webserver Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:54:36.496: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:54:36.498: INFO: ExecWithOptions: Clientset creation - Dec 14 16:54:36.498: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-8153/pods/test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+%27http%3A%2F%2F10.233.66.146%3A9080%2Fdial%3Frequest%3Dhostname%26protocol%3Dudp%26host%3D10.233.66.145%26port%3D8081%26tries%3D1%27&container=webserver&container=webserver&stderr=true&stdout=true) - Dec 14 16:54:36.645: INFO: Waiting for responses: map[] - Dec 14 16:54:36.645: INFO: reached 10.233.66.145 after 0/1 tries - Dec 14 16:54:36.645: INFO: Going to retry 0 out of 3 pods.... - [AfterEach] [sig-network] Networking + STEP: Creating a kubernetes client 02/25/23 11:23:03.689 + Feb 25 11:23:03.689: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 11:23:03.693 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:23:03.717 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:23:03.722 + [BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 + [It] should provide podname only [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:52 + STEP: Creating a pod to test downward API volume plugin 02/25/23 11:23:03.726 + Feb 25 11:23:03.750: INFO: Waiting up to 5m0s for pod "downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128" in namespace "downward-api-8575" to be "Succeeded or Failed" + Feb 25 11:23:03.762: INFO: Pod "downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128": Phase="Pending", Reason="", readiness=false. Elapsed: 12.349531ms + Feb 25 11:23:05.770: INFO: Pod "downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019790611s + Feb 25 11:23:07.768: INFO: Pod "downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018598863s + STEP: Saw pod success 02/25/23 11:23:07.769 + Feb 25 11:23:07.769: INFO: Pod "downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128" satisfied condition "Succeeded or Failed" + Feb 25 11:23:07.774: INFO: Trying to get logs from node diek3dua9gay-2 pod downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128 container client-container: + STEP: delete the pod 02/25/23 11:23:07.795 + Feb 25 11:23:07.815: INFO: Waiting for pod downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128 to disappear + Feb 25 11:23:07.820: INFO: Pod downwardapi-volume-60b72d1e-f284-4792-9445-bc04e61f3128 no longer exists + [AfterEach] [sig-storage] Downward API volume test/e2e/framework/framework.go:187 - Dec 14 16:54:36.645: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "pod-network-test-8153" for this suite. 12/14/22 16:54:36.653 + Feb 25 11:23:07.821: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-8575" for this suite. 02/25/23 11:23:07.827 << End Captured GinkgoWriter Output ------------------------------ SSSSSSSSS ------------------------------ -[sig-api-machinery] Watchers - should observe add, update, and delete watch notifications on configmaps [Conformance] - test/e2e/apimachinery/watch.go:60 -[BeforeEach] [sig-api-machinery] Watchers +[sig-storage] Projected downwardAPI + should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:67 +[BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:54:36.664 -Dec 14 16:54:36.664: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename watch 12/14/22 16:54:36.667 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:54:36.693 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:54:36.696 -[It] should observe add, update, and delete watch notifications on configmaps [Conformance] - test/e2e/apimachinery/watch.go:60 -STEP: creating a watch on configmaps with label A 12/14/22 16:54:36.699 -STEP: creating a watch on configmaps with label B 12/14/22 16:54:36.701 -STEP: creating a watch on configmaps with label A or B 12/14/22 16:54:36.703 -STEP: creating a configmap with label A and ensuring the correct watchers observe the notification 12/14/22 16:54:36.704 -Dec 14 16:54:36.710: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27141 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} -Dec 14 16:54:36.710: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27141 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} -STEP: modifying configmap A and ensuring the correct watchers observe the notification 12/14/22 16:54:36.711 -Dec 14 16:54:36.727: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27142 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} -Dec 14 16:54:36.728: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27142 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} -STEP: modifying configmap A again and ensuring the correct watchers observe the notification 12/14/22 16:54:36.728 -Dec 14 16:54:36.737: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27143 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} -Dec 14 16:54:36.737: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27143 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} -STEP: deleting configmap A and ensuring the correct watchers observe the notification 12/14/22 16:54:36.737 -Dec 14 16:54:36.746: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27144 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} -Dec 14 16:54:36.746: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27144 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} -STEP: creating a configmap with label B and ensuring the correct watchers observe the notification 12/14/22 16:54:36.746 -Dec 14 16:54:36.752: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-3620 30e4c08f-5dca-4a58-9af5-9700edc0f950 27145 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} -Dec 14 16:54:36.752: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-3620 30e4c08f-5dca-4a58-9af5-9700edc0f950 27145 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} -STEP: deleting configmap B and ensuring the correct watchers observe the notification 12/14/22 16:54:46.753 -Dec 14 16:54:46.768: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-3620 30e4c08f-5dca-4a58-9af5-9700edc0f950 27198 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} -Dec 14 16:54:46.769: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-3620 30e4c08f-5dca-4a58-9af5-9700edc0f950 27198 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} -[AfterEach] [sig-api-machinery] Watchers +STEP: Creating a kubernetes client 02/25/23 11:23:07.841 +Feb 25 11:23:07.842: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 11:23:07.845 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:23:07.864 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:23:07.87 +[BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 +[It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:67 +STEP: Creating a pod to test downward API volume plugin 02/25/23 11:23:07.874 +Feb 25 11:23:07.887: INFO: Waiting up to 5m0s for pod "downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5" in namespace "projected-2138" to be "Succeeded or Failed" +Feb 25 11:23:07.899: INFO: Pod "downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5": Phase="Pending", Reason="", readiness=false. Elapsed: 11.509277ms +Feb 25 11:23:09.912: INFO: Pod "downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024191606s +Feb 25 11:23:11.905: INFO: Pod "downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.017605759s +STEP: Saw pod success 02/25/23 11:23:11.905 +Feb 25 11:23:11.905: INFO: Pod "downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5" satisfied condition "Succeeded or Failed" +Feb 25 11:23:11.917: INFO: Trying to get logs from node diek3dua9gay-2 pod downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5 container client-container: +STEP: delete the pod 02/25/23 11:23:11.929 +Feb 25 11:23:11.952: INFO: Waiting for pod downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5 to disappear +Feb 25 11:23:11.960: INFO: Pod downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5 no longer exists +[AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 -Dec 14 16:54:56.769: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "watch-3620" for this suite. 12/14/22 16:54:56.78 -{"msg":"PASSED [sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance]","completed":281,"skipped":5325,"failed":0} +Feb 25 11:23:11.960: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-2138" for this suite. 02/25/23 11:23:11.969 +{"msg":"PASSED [sig-storage] Projected downwardAPI should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance]","completed":278,"skipped":5287,"failed":0} ------------------------------ -• [SLOW TEST] [20.125 seconds] -[sig-api-machinery] Watchers -test/e2e/apimachinery/framework.go:23 - should observe add, update, and delete watch notifications on configmaps [Conformance] - test/e2e/apimachinery/watch.go:60 +• [4.146 seconds] +[sig-storage] Projected downwardAPI +test/e2e/common/storage/framework.go:23 + should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:67 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Watchers + [BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:54:36.664 - Dec 14 16:54:36.664: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename watch 12/14/22 16:54:36.667 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:54:36.693 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:54:36.696 - [It] should observe add, update, and delete watch notifications on configmaps [Conformance] - test/e2e/apimachinery/watch.go:60 - STEP: creating a watch on configmaps with label A 12/14/22 16:54:36.699 - STEP: creating a watch on configmaps with label B 12/14/22 16:54:36.701 - STEP: creating a watch on configmaps with label A or B 12/14/22 16:54:36.703 - STEP: creating a configmap with label A and ensuring the correct watchers observe the notification 12/14/22 16:54:36.704 - Dec 14 16:54:36.710: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27141 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} - Dec 14 16:54:36.710: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27141 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} - STEP: modifying configmap A and ensuring the correct watchers observe the notification 12/14/22 16:54:36.711 - Dec 14 16:54:36.727: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27142 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} - Dec 14 16:54:36.728: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27142 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} - STEP: modifying configmap A again and ensuring the correct watchers observe the notification 12/14/22 16:54:36.728 - Dec 14 16:54:36.737: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27143 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} - Dec 14 16:54:36.737: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27143 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} - STEP: deleting configmap A and ensuring the correct watchers observe the notification 12/14/22 16:54:36.737 - Dec 14 16:54:36.746: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27144 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} - Dec 14 16:54:36.746: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-3620 6ababa58-70b8-4f00-b14a-98b39916080c 27144 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} - STEP: creating a configmap with label B and ensuring the correct watchers observe the notification 12/14/22 16:54:36.746 - Dec 14 16:54:36.752: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-3620 30e4c08f-5dca-4a58-9af5-9700edc0f950 27145 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} - Dec 14 16:54:36.752: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-3620 30e4c08f-5dca-4a58-9af5-9700edc0f950 27145 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} - STEP: deleting configmap B and ensuring the correct watchers observe the notification 12/14/22 16:54:46.753 - Dec 14 16:54:46.768: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-3620 30e4c08f-5dca-4a58-9af5-9700edc0f950 27198 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} - Dec 14 16:54:46.769: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-3620 30e4c08f-5dca-4a58-9af5-9700edc0f950 27198 0 2022-12-14 16:54:36 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2022-12-14 16:54:36 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} - [AfterEach] [sig-api-machinery] Watchers + STEP: Creating a kubernetes client 02/25/23 11:23:07.841 + Feb 25 11:23:07.842: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 11:23:07.845 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:23:07.864 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:23:07.87 + [BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 + [It] should set DefaultMode on files [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:67 + STEP: Creating a pod to test downward API volume plugin 02/25/23 11:23:07.874 + Feb 25 11:23:07.887: INFO: Waiting up to 5m0s for pod "downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5" in namespace "projected-2138" to be "Succeeded or Failed" + Feb 25 11:23:07.899: INFO: Pod "downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5": Phase="Pending", Reason="", readiness=false. Elapsed: 11.509277ms + Feb 25 11:23:09.912: INFO: Pod "downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5": Phase="Pending", Reason="", readiness=false. Elapsed: 2.024191606s + Feb 25 11:23:11.905: INFO: Pod "downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.017605759s + STEP: Saw pod success 02/25/23 11:23:11.905 + Feb 25 11:23:11.905: INFO: Pod "downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5" satisfied condition "Succeeded or Failed" + Feb 25 11:23:11.917: INFO: Trying to get logs from node diek3dua9gay-2 pod downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5 container client-container: + STEP: delete the pod 02/25/23 11:23:11.929 + Feb 25 11:23:11.952: INFO: Waiting for pod downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5 to disappear + Feb 25 11:23:11.960: INFO: Pod downwardapi-volume-072c2970-5b0a-462f-b9fc-75948f2141c5 no longer exists + [AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 - Dec 14 16:54:56.769: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "watch-3620" for this suite. 12/14/22 16:54:56.78 + Feb 25 11:23:11.960: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-2138" for this suite. 02/25/23 11:23:11.969 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-architecture] Conformance Tests - should have at least two untainted nodes [Conformance] - test/e2e/architecture/conformance.go:38 -[BeforeEach] [sig-architecture] Conformance Tests +[sig-storage] EmptyDir volumes + should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:146 +[BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:54:56.803 -Dec 14 16:54:56.804: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename conformance-tests 12/14/22 16:54:56.807 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:54:56.833 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:54:56.836 -[It] should have at least two untainted nodes [Conformance] - test/e2e/architecture/conformance.go:38 -STEP: Getting node addresses 12/14/22 16:54:56.839 -Dec 14 16:54:56.840: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable -[AfterEach] [sig-architecture] Conformance Tests +STEP: Creating a kubernetes client 02/25/23 11:23:11.994 +Feb 25 11:23:11.994: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir 02/25/23 11:23:12 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:23:12.02 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:23:12.027 +[It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:146 +STEP: Creating a pod to test emptydir 0777 on tmpfs 02/25/23 11:23:12.031 +Feb 25 11:23:12.044: INFO: Waiting up to 5m0s for pod "pod-9b2633f3-d31b-483f-8962-1f3cc6772033" in namespace "emptydir-545" to be "Succeeded or Failed" +Feb 25 11:23:12.050: INFO: Pod "pod-9b2633f3-d31b-483f-8962-1f3cc6772033": Phase="Pending", Reason="", readiness=false. Elapsed: 6.185114ms +Feb 25 11:23:14.057: INFO: Pod "pod-9b2633f3-d31b-483f-8962-1f3cc6772033": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01294262s +Feb 25 11:23:16.059: INFO: Pod "pod-9b2633f3-d31b-483f-8962-1f3cc6772033": Phase="Pending", Reason="", readiness=false. Elapsed: 4.015173165s +Feb 25 11:23:18.059: INFO: Pod "pod-9b2633f3-d31b-483f-8962-1f3cc6772033": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.01456365s +STEP: Saw pod success 02/25/23 11:23:18.059 +Feb 25 11:23:18.059: INFO: Pod "pod-9b2633f3-d31b-483f-8962-1f3cc6772033" satisfied condition "Succeeded or Failed" +Feb 25 11:23:18.065: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-9b2633f3-d31b-483f-8962-1f3cc6772033 container test-container: +STEP: delete the pod 02/25/23 11:23:18.202 +Feb 25 11:23:18.225: INFO: Waiting for pod pod-9b2633f3-d31b-483f-8962-1f3cc6772033 to disappear +Feb 25 11:23:18.229: INFO: Pod pod-9b2633f3-d31b-483f-8962-1f3cc6772033 no longer exists +[AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 -Dec 14 16:54:56.848: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "conformance-tests-4934" for this suite. 12/14/22 16:54:56.852 -{"msg":"PASSED [sig-architecture] Conformance Tests should have at least two untainted nodes [Conformance]","completed":282,"skipped":5374,"failed":0} +Feb 25 11:23:18.229: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-545" for this suite. 02/25/23 11:23:18.236 +{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","completed":279,"skipped":5312,"failed":0} ------------------------------ -• [0.057 seconds] -[sig-architecture] Conformance Tests -test/e2e/architecture/framework.go:23 - should have at least two untainted nodes [Conformance] - test/e2e/architecture/conformance.go:38 +• [SLOW TEST] [6.255 seconds] +[sig-storage] EmptyDir volumes +test/e2e/common/storage/framework.go:23 + should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:146 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-architecture] Conformance Tests + [BeforeEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:54:56.803 - Dec 14 16:54:56.804: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename conformance-tests 12/14/22 16:54:56.807 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:54:56.833 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:54:56.836 - [It] should have at least two untainted nodes [Conformance] - test/e2e/architecture/conformance.go:38 - STEP: Getting node addresses 12/14/22 16:54:56.839 - Dec 14 16:54:56.840: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable - [AfterEach] [sig-architecture] Conformance Tests + STEP: Creating a kubernetes client 02/25/23 11:23:11.994 + Feb 25 11:23:11.994: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir 02/25/23 11:23:12 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:23:12.02 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:23:12.027 + [It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/empty_dir.go:146 + STEP: Creating a pod to test emptydir 0777 on tmpfs 02/25/23 11:23:12.031 + Feb 25 11:23:12.044: INFO: Waiting up to 5m0s for pod "pod-9b2633f3-d31b-483f-8962-1f3cc6772033" in namespace "emptydir-545" to be "Succeeded or Failed" + Feb 25 11:23:12.050: INFO: Pod "pod-9b2633f3-d31b-483f-8962-1f3cc6772033": Phase="Pending", Reason="", readiness=false. Elapsed: 6.185114ms + Feb 25 11:23:14.057: INFO: Pod "pod-9b2633f3-d31b-483f-8962-1f3cc6772033": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01294262s + Feb 25 11:23:16.059: INFO: Pod "pod-9b2633f3-d31b-483f-8962-1f3cc6772033": Phase="Pending", Reason="", readiness=false. Elapsed: 4.015173165s + Feb 25 11:23:18.059: INFO: Pod "pod-9b2633f3-d31b-483f-8962-1f3cc6772033": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.01456365s + STEP: Saw pod success 02/25/23 11:23:18.059 + Feb 25 11:23:18.059: INFO: Pod "pod-9b2633f3-d31b-483f-8962-1f3cc6772033" satisfied condition "Succeeded or Failed" + Feb 25 11:23:18.065: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-9b2633f3-d31b-483f-8962-1f3cc6772033 container test-container: + STEP: delete the pod 02/25/23 11:23:18.202 + Feb 25 11:23:18.225: INFO: Waiting for pod pod-9b2633f3-d31b-483f-8962-1f3cc6772033 to disappear + Feb 25 11:23:18.229: INFO: Pod pod-9b2633f3-d31b-483f-8962-1f3cc6772033 no longer exists + [AfterEach] [sig-storage] EmptyDir volumes test/e2e/framework/framework.go:187 - Dec 14 16:54:56.848: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "conformance-tests-4934" for this suite. 12/14/22 16:54:56.852 + Feb 25 11:23:18.229: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-545" for this suite. 02/25/23 11:23:18.236 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSS ------------------------------ -[sig-network] Services - should have session affinity work for NodePort service [LinuxOnly] [Conformance] - test/e2e/network/service.go:2204 -[BeforeEach] [sig-network] Services +[sig-api-machinery] Watchers + should be able to start watching from a specific resource version [Conformance] + test/e2e/apimachinery/watch.go:142 +[BeforeEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:54:56.865 -Dec 14 16:54:56.865: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 16:54:56.867 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:54:56.893 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:54:56.898 -[BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 -[It] should have session affinity work for NodePort service [LinuxOnly] [Conformance] - test/e2e/network/service.go:2204 -STEP: creating service in namespace services-1800 12/14/22 16:54:56.901 -STEP: creating service affinity-nodeport in namespace services-1800 12/14/22 16:54:56.901 -STEP: creating replication controller affinity-nodeport in namespace services-1800 12/14/22 16:54:56.933 -I1214 16:54:56.945332 14 runners.go:193] Created replication controller with name: affinity-nodeport, namespace: services-1800, replica count: 3 -I1214 16:54:59.997073 14 runners.go:193] affinity-nodeport Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -Dec 14 16:55:00.017: INFO: Creating new exec pod -Dec 14 16:55:00.038: INFO: Waiting up to 5m0s for pod "execpod-affinityqqpp9" in namespace "services-1800" to be "running" -Dec 14 16:55:00.048: INFO: Pod "execpod-affinityqqpp9": Phase="Pending", Reason="", readiness=false. Elapsed: 8.919726ms -Dec 14 16:55:02.058: INFO: Pod "execpod-affinityqqpp9": Phase="Running", Reason="", readiness=true. Elapsed: 2.019556659s -Dec 14 16:55:02.058: INFO: Pod "execpod-affinityqqpp9" satisfied condition "running" -Dec 14 16:55:03.065: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1800 exec execpod-affinityqqpp9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-nodeport 80' -Dec 14 16:55:03.296: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-nodeport 80\nConnection to affinity-nodeport 80 port [tcp/http] succeeded!\n" -Dec 14 16:55:03.297: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:55:03.297: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1800 exec execpod-affinityqqpp9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.43.198 80' -Dec 14 16:55:03.477: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.43.198 80\nConnection to 10.233.43.198 80 port [tcp/http] succeeded!\n" -Dec 14 16:55:03.477: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:55:03.478: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1800 exec execpod-affinityqqpp9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.244 32260' -Dec 14 16:55:03.671: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.244 32260\nConnection to 192.168.121.244 32260 port [tcp/*] succeeded!\n" -Dec 14 16:55:03.671: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:55:03.672: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1800 exec execpod-affinityqqpp9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.197 32260' -Dec 14 16:55:04.296: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.197 32260\nConnection to 192.168.121.197 32260 port [tcp/*] succeeded!\n" -Dec 14 16:55:04.296: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:55:04.297: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1800 exec execpod-affinityqqpp9 -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.249:32260/ ; done' -Dec 14 16:55:04.844: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n" -Dec 14 16:55:04.844: INFO: stdout: "\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv" -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv -Dec 14 16:55:04.844: INFO: Cleaning up the exec pod -STEP: deleting ReplicationController affinity-nodeport in namespace services-1800, will wait for the garbage collector to delete the pods 12/14/22 16:55:04.868 -Dec 14 16:55:04.938: INFO: Deleting ReplicationController affinity-nodeport took: 10.926666ms -Dec 14 16:55:05.038: INFO: Terminating ReplicationController affinity-nodeport pods took: 100.455123ms -[AfterEach] [sig-network] Services +STEP: Creating a kubernetes client 02/25/23 11:23:18.253 +Feb 25 11:23:18.253: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename watch 02/25/23 11:23:18.256 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:23:18.283 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:23:18.287 +[It] should be able to start watching from a specific resource version [Conformance] + test/e2e/apimachinery/watch.go:142 +STEP: creating a new configmap 02/25/23 11:23:18.29 +STEP: modifying the configmap once 02/25/23 11:23:18.297 +STEP: modifying the configmap a second time 02/25/23 11:23:18.309 +STEP: deleting the configmap 02/25/23 11:23:18.328 +STEP: creating a watch on configmaps from the resource version returned by the first update 02/25/23 11:23:18.337 +STEP: Expecting to observe notifications for all changes to the configmap after the first update 02/25/23 11:23:18.339 +Feb 25 11:23:18.339: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-6099 7dc20f4e-3a49-4958-99fb-75a5465e0efc 41262 0 2023-02-25 11:23:18 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] [{e2e.test Update v1 2023-02-25 11:23:18 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} +Feb 25 11:23:18.340: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-6099 7dc20f4e-3a49-4958-99fb-75a5465e0efc 41263 0 2023-02-25 11:23:18 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] [{e2e.test Update v1 2023-02-25 11:23:18 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} +[AfterEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:187 -Dec 14 16:55:07.479: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-1800" for this suite. 12/14/22 16:55:07.486 -[AfterEach] [sig-network] Services - test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should have session affinity work for NodePort service [LinuxOnly] [Conformance]","completed":283,"skipped":5409,"failed":0} +Feb 25 11:23:18.340: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "watch-6099" for this suite. 02/25/23 11:23:18.358 +{"msg":"PASSED [sig-api-machinery] Watchers should be able to start watching from a specific resource version [Conformance]","completed":280,"skipped":5319,"failed":0} ------------------------------ -• [SLOW TEST] [10.629 seconds] -[sig-network] Services -test/e2e/network/common/framework.go:23 - should have session affinity work for NodePort service [LinuxOnly] [Conformance] - test/e2e/network/service.go:2204 +• [0.118 seconds] +[sig-api-machinery] Watchers +test/e2e/apimachinery/framework.go:23 + should be able to start watching from a specific resource version [Conformance] + test/e2e/apimachinery/watch.go:142 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Services + [BeforeEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:54:56.865 - Dec 14 16:54:56.865: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 16:54:56.867 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:54:56.893 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:54:56.898 - [BeforeEach] [sig-network] Services - test/e2e/network/service.go:758 - [It] should have session affinity work for NodePort service [LinuxOnly] [Conformance] - test/e2e/network/service.go:2204 - STEP: creating service in namespace services-1800 12/14/22 16:54:56.901 - STEP: creating service affinity-nodeport in namespace services-1800 12/14/22 16:54:56.901 - STEP: creating replication controller affinity-nodeport in namespace services-1800 12/14/22 16:54:56.933 - I1214 16:54:56.945332 14 runners.go:193] Created replication controller with name: affinity-nodeport, namespace: services-1800, replica count: 3 - I1214 16:54:59.997073 14 runners.go:193] affinity-nodeport Pods: 3 out of 3 created, 3 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - Dec 14 16:55:00.017: INFO: Creating new exec pod - Dec 14 16:55:00.038: INFO: Waiting up to 5m0s for pod "execpod-affinityqqpp9" in namespace "services-1800" to be "running" - Dec 14 16:55:00.048: INFO: Pod "execpod-affinityqqpp9": Phase="Pending", Reason="", readiness=false. Elapsed: 8.919726ms - Dec 14 16:55:02.058: INFO: Pod "execpod-affinityqqpp9": Phase="Running", Reason="", readiness=true. Elapsed: 2.019556659s - Dec 14 16:55:02.058: INFO: Pod "execpod-affinityqqpp9" satisfied condition "running" - Dec 14 16:55:03.065: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1800 exec execpod-affinityqqpp9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 affinity-nodeport 80' - Dec 14 16:55:03.296: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 affinity-nodeport 80\nConnection to affinity-nodeport 80 port [tcp/http] succeeded!\n" - Dec 14 16:55:03.297: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:55:03.297: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1800 exec execpod-affinityqqpp9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.43.198 80' - Dec 14 16:55:03.477: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.43.198 80\nConnection to 10.233.43.198 80 port [tcp/http] succeeded!\n" - Dec 14 16:55:03.477: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:55:03.478: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1800 exec execpod-affinityqqpp9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.244 32260' - Dec 14 16:55:03.671: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.244 32260\nConnection to 192.168.121.244 32260 port [tcp/*] succeeded!\n" - Dec 14 16:55:03.671: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:55:03.672: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1800 exec execpod-affinityqqpp9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.197 32260' - Dec 14 16:55:04.296: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.197 32260\nConnection to 192.168.121.197 32260 port [tcp/*] succeeded!\n" - Dec 14 16:55:04.296: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:55:04.297: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-1800 exec execpod-affinityqqpp9 -- /bin/sh -x -c for i in $(seq 0 15); do echo; curl -q -s --connect-timeout 2 http://192.168.121.249:32260/ ; done' - Dec 14 16:55:04.844: INFO: stderr: "+ seq 0 15\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n+ echo\n+ curl -q -s --connect-timeout 2 http://192.168.121.249:32260/\n" - Dec 14 16:55:04.844: INFO: stdout: "\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv\naffinity-nodeport-7xvvv" - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Received response from host: affinity-nodeport-7xvvv - Dec 14 16:55:04.844: INFO: Cleaning up the exec pod - STEP: deleting ReplicationController affinity-nodeport in namespace services-1800, will wait for the garbage collector to delete the pods 12/14/22 16:55:04.868 - Dec 14 16:55:04.938: INFO: Deleting ReplicationController affinity-nodeport took: 10.926666ms - Dec 14 16:55:05.038: INFO: Terminating ReplicationController affinity-nodeport pods took: 100.455123ms - [AfterEach] [sig-network] Services + STEP: Creating a kubernetes client 02/25/23 11:23:18.253 + Feb 25 11:23:18.253: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename watch 02/25/23 11:23:18.256 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:23:18.283 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:23:18.287 + [It] should be able to start watching from a specific resource version [Conformance] + test/e2e/apimachinery/watch.go:142 + STEP: creating a new configmap 02/25/23 11:23:18.29 + STEP: modifying the configmap once 02/25/23 11:23:18.297 + STEP: modifying the configmap a second time 02/25/23 11:23:18.309 + STEP: deleting the configmap 02/25/23 11:23:18.328 + STEP: creating a watch on configmaps from the resource version returned by the first update 02/25/23 11:23:18.337 + STEP: Expecting to observe notifications for all changes to the configmap after the first update 02/25/23 11:23:18.339 + Feb 25 11:23:18.339: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-6099 7dc20f4e-3a49-4958-99fb-75a5465e0efc 41262 0 2023-02-25 11:23:18 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] [{e2e.test Update v1 2023-02-25 11:23:18 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} + Feb 25 11:23:18.340: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-resource-version watch-6099 7dc20f4e-3a49-4958-99fb-75a5465e0efc 41263 0 2023-02-25 11:23:18 +0000 UTC map[watch-this-configmap:from-resource-version] map[] [] [] [{e2e.test Update v1 2023-02-25 11:23:18 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} + [AfterEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:187 - Dec 14 16:55:07.479: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-1800" for this suite. 12/14/22 16:55:07.486 - [AfterEach] [sig-network] Services - test/e2e/network/service.go:762 + Feb 25 11:23:18.340: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "watch-6099" for this suite. 02/25/23 11:23:18.358 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSS +S ------------------------------ -[sig-apps] ReplicationController - should surface a failure condition on a common issue like exceeded quota [Conformance] - test/e2e/apps/rc.go:82 -[BeforeEach] [sig-apps] ReplicationController +[sig-storage] Projected downwardAPI + should provide container's cpu limit [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:192 +[BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:55:07.502 -Dec 14 16:55:07.502: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename replication-controller 12/14/22 16:55:07.504 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:07.532 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:07.535 -[BeforeEach] [sig-apps] ReplicationController - test/e2e/apps/rc.go:56 -[It] should surface a failure condition on a common issue like exceeded quota [Conformance] - test/e2e/apps/rc.go:82 -Dec 14 16:55:07.539: INFO: Creating quota "condition-test" that allows only two pods to run in the current namespace -STEP: Creating rc "condition-test" that asks for more than the allowed pod quota 12/14/22 16:55:08.558 -STEP: Checking rc "condition-test" has the desired failure condition set 12/14/22 16:55:08.566 -STEP: Scaling down rc "condition-test" to satisfy pod quota 12/14/22 16:55:09.581 -Dec 14 16:55:09.595: INFO: Updating replication controller "condition-test" -STEP: Checking rc "condition-test" has no failure condition set 12/14/22 16:55:09.595 -[AfterEach] [sig-apps] ReplicationController +STEP: Creating a kubernetes client 02/25/23 11:23:18.374 +Feb 25 11:23:18.374: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 11:23:18.376 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:23:18.407 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:23:18.412 +[BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 +[It] should provide container's cpu limit [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:192 +STEP: Creating a pod to test downward API volume plugin 02/25/23 11:23:18.416 +Feb 25 11:23:18.436: INFO: Waiting up to 5m0s for pod "downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8" in namespace "projected-6213" to be "Succeeded or Failed" +Feb 25 11:23:18.443: INFO: Pod "downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8": Phase="Pending", Reason="", readiness=false. Elapsed: 7.26599ms +Feb 25 11:23:20.450: INFO: Pod "downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014403268s +Feb 25 11:23:22.449: INFO: Pod "downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013147475s +STEP: Saw pod success 02/25/23 11:23:22.449 +Feb 25 11:23:22.449: INFO: Pod "downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8" satisfied condition "Succeeded or Failed" +Feb 25 11:23:22.456: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8 container client-container: +STEP: delete the pod 02/25/23 11:23:22.465 +Feb 25 11:23:22.485: INFO: Waiting for pod downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8 to disappear +Feb 25 11:23:22.490: INFO: Pod downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8 no longer exists +[AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 -Dec 14 16:55:10.609: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "replication-controller-71" for this suite. 12/14/22 16:55:10.614 -{"msg":"PASSED [sig-apps] ReplicationController should surface a failure condition on a common issue like exceeded quota [Conformance]","completed":284,"skipped":5434,"failed":0} +Feb 25 11:23:22.490: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-6213" for this suite. 02/25/23 11:23:22.496 +{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance]","completed":281,"skipped":5320,"failed":0} ------------------------------ -• [3.121 seconds] -[sig-apps] ReplicationController -test/e2e/apps/framework.go:23 - should surface a failure condition on a common issue like exceeded quota [Conformance] - test/e2e/apps/rc.go:82 +• [4.132 seconds] +[sig-storage] Projected downwardAPI +test/e2e/common/storage/framework.go:23 + should provide container's cpu limit [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:192 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] ReplicationController + [BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:55:07.502 - Dec 14 16:55:07.502: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename replication-controller 12/14/22 16:55:07.504 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:07.532 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:07.535 - [BeforeEach] [sig-apps] ReplicationController - test/e2e/apps/rc.go:56 - [It] should surface a failure condition on a common issue like exceeded quota [Conformance] - test/e2e/apps/rc.go:82 - Dec 14 16:55:07.539: INFO: Creating quota "condition-test" that allows only two pods to run in the current namespace - STEP: Creating rc "condition-test" that asks for more than the allowed pod quota 12/14/22 16:55:08.558 - STEP: Checking rc "condition-test" has the desired failure condition set 12/14/22 16:55:08.566 - STEP: Scaling down rc "condition-test" to satisfy pod quota 12/14/22 16:55:09.581 - Dec 14 16:55:09.595: INFO: Updating replication controller "condition-test" - STEP: Checking rc "condition-test" has no failure condition set 12/14/22 16:55:09.595 - [AfterEach] [sig-apps] ReplicationController + STEP: Creating a kubernetes client 02/25/23 11:23:18.374 + Feb 25 11:23:18.374: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 11:23:18.376 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:23:18.407 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:23:18.412 + [BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 + [It] should provide container's cpu limit [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:192 + STEP: Creating a pod to test downward API volume plugin 02/25/23 11:23:18.416 + Feb 25 11:23:18.436: INFO: Waiting up to 5m0s for pod "downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8" in namespace "projected-6213" to be "Succeeded or Failed" + Feb 25 11:23:18.443: INFO: Pod "downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8": Phase="Pending", Reason="", readiness=false. Elapsed: 7.26599ms + Feb 25 11:23:20.450: INFO: Pod "downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014403268s + Feb 25 11:23:22.449: INFO: Pod "downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013147475s + STEP: Saw pod success 02/25/23 11:23:22.449 + Feb 25 11:23:22.449: INFO: Pod "downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8" satisfied condition "Succeeded or Failed" + Feb 25 11:23:22.456: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8 container client-container: + STEP: delete the pod 02/25/23 11:23:22.465 + Feb 25 11:23:22.485: INFO: Waiting for pod downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8 to disappear + Feb 25 11:23:22.490: INFO: Pod downwardapi-volume-8181d86c-3fee-4287-a524-991cce1cf7c8 no longer exists + [AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 - Dec 14 16:55:10.609: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "replication-controller-71" for this suite. 12/14/22 16:55:10.614 + Feb 25 11:23:22.490: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-6213" for this suite. 02/25/23 11:23:22.496 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSS ------------------------------ -[sig-api-machinery] ResourceQuota - should create a ResourceQuota and capture the life of a replica set. [Conformance] - test/e2e/apimachinery/resource_quota.go:438 -[BeforeEach] [sig-api-machinery] ResourceQuota +[sig-storage] Projected configMap + optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:173 +[BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:55:10.633 -Dec 14 16:55:10.633: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename resourcequota 12/14/22 16:55:10.636 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:10.661 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:10.665 -[It] should create a ResourceQuota and capture the life of a replica set. [Conformance] - test/e2e/apimachinery/resource_quota.go:438 -STEP: Counting existing ResourceQuota 12/14/22 16:55:10.668 -STEP: Creating a ResourceQuota 12/14/22 16:55:15.675 -STEP: Ensuring resource quota status is calculated 12/14/22 16:55:15.683 -STEP: Creating a ReplicaSet 12/14/22 16:55:17.694 -STEP: Ensuring resource quota status captures replicaset creation 12/14/22 16:55:17.718 -STEP: Deleting a ReplicaSet 12/14/22 16:55:19.727 -STEP: Ensuring resource quota status released usage 12/14/22 16:55:19.736 -[AfterEach] [sig-api-machinery] ResourceQuota +STEP: Creating a kubernetes client 02/25/23 11:23:22.513 +Feb 25 11:23:22.513: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 11:23:22.519 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:23:22.548 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:23:22.555 +[It] optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:173 +STEP: Creating configMap with name cm-test-opt-del-e3a9a655-c405-4ed7-a320-3c466fa0061f 02/25/23 11:23:22.565 +STEP: Creating configMap with name cm-test-opt-upd-ee90c78e-12b9-469a-b8e1-91fa86ef2047 02/25/23 11:23:22.571 +STEP: Creating the pod 02/25/23 11:23:22.58 +Feb 25 11:23:22.594: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f" in namespace "projected-391" to be "running and ready" +Feb 25 11:23:22.602: INFO: Pod "pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f": Phase="Pending", Reason="", readiness=false. Elapsed: 7.735698ms +Feb 25 11:23:22.602: INFO: The phase of Pod pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:23:24.613: INFO: Pod "pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018630356s +Feb 25 11:23:24.613: INFO: The phase of Pod pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:23:26.610: INFO: Pod "pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f": Phase="Running", Reason="", readiness=true. Elapsed: 4.015999647s +Feb 25 11:23:26.610: INFO: The phase of Pod pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f is Running (Ready = true) +Feb 25 11:23:26.610: INFO: Pod "pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f" satisfied condition "running and ready" +STEP: Deleting configmap cm-test-opt-del-e3a9a655-c405-4ed7-a320-3c466fa0061f 02/25/23 11:23:26.646 +STEP: Updating configmap cm-test-opt-upd-ee90c78e-12b9-469a-b8e1-91fa86ef2047 02/25/23 11:23:26.657 +STEP: Creating configMap with name cm-test-opt-create-b6d0b565-7ec4-4d86-acaa-f49f6eba659d 02/25/23 11:23:26.665 +STEP: waiting to observe update in volume 02/25/23 11:23:26.672 +[AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 -Dec 14 16:55:21.747: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-9821" for this suite. 12/14/22 16:55:21.764 -{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replica set. [Conformance]","completed":285,"skipped":5465,"failed":0} +Feb 25 11:24:53.557: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-391" for this suite. 02/25/23 11:24:53.576 +{"msg":"PASSED [sig-storage] Projected configMap optional updates should be reflected in volume [NodeConformance] [Conformance]","completed":282,"skipped":5325,"failed":0} ------------------------------ -• [SLOW TEST] [11.152 seconds] -[sig-api-machinery] ResourceQuota -test/e2e/apimachinery/framework.go:23 - should create a ResourceQuota and capture the life of a replica set. [Conformance] - test/e2e/apimachinery/resource_quota.go:438 +• [SLOW TEST] [91.091 seconds] +[sig-storage] Projected configMap +test/e2e/common/storage/framework.go:23 + optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:173 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] ResourceQuota + [BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:55:10.633 - Dec 14 16:55:10.633: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename resourcequota 12/14/22 16:55:10.636 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:10.661 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:10.665 - [It] should create a ResourceQuota and capture the life of a replica set. [Conformance] - test/e2e/apimachinery/resource_quota.go:438 - STEP: Counting existing ResourceQuota 12/14/22 16:55:10.668 - STEP: Creating a ResourceQuota 12/14/22 16:55:15.675 - STEP: Ensuring resource quota status is calculated 12/14/22 16:55:15.683 - STEP: Creating a ReplicaSet 12/14/22 16:55:17.694 - STEP: Ensuring resource quota status captures replicaset creation 12/14/22 16:55:17.718 - STEP: Deleting a ReplicaSet 12/14/22 16:55:19.727 - STEP: Ensuring resource quota status released usage 12/14/22 16:55:19.736 - [AfterEach] [sig-api-machinery] ResourceQuota + STEP: Creating a kubernetes client 02/25/23 11:23:22.513 + Feb 25 11:23:22.513: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 11:23:22.519 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:23:22.548 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:23:22.555 + [It] optional updates should be reflected in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:173 + STEP: Creating configMap with name cm-test-opt-del-e3a9a655-c405-4ed7-a320-3c466fa0061f 02/25/23 11:23:22.565 + STEP: Creating configMap with name cm-test-opt-upd-ee90c78e-12b9-469a-b8e1-91fa86ef2047 02/25/23 11:23:22.571 + STEP: Creating the pod 02/25/23 11:23:22.58 + Feb 25 11:23:22.594: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f" in namespace "projected-391" to be "running and ready" + Feb 25 11:23:22.602: INFO: Pod "pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f": Phase="Pending", Reason="", readiness=false. Elapsed: 7.735698ms + Feb 25 11:23:22.602: INFO: The phase of Pod pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:23:24.613: INFO: Pod "pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.018630356s + Feb 25 11:23:24.613: INFO: The phase of Pod pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:23:26.610: INFO: Pod "pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f": Phase="Running", Reason="", readiness=true. Elapsed: 4.015999647s + Feb 25 11:23:26.610: INFO: The phase of Pod pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f is Running (Ready = true) + Feb 25 11:23:26.610: INFO: Pod "pod-projected-configmaps-c2850517-5b99-4ecd-9e35-5d463c413b6f" satisfied condition "running and ready" + STEP: Deleting configmap cm-test-opt-del-e3a9a655-c405-4ed7-a320-3c466fa0061f 02/25/23 11:23:26.646 + STEP: Updating configmap cm-test-opt-upd-ee90c78e-12b9-469a-b8e1-91fa86ef2047 02/25/23 11:23:26.657 + STEP: Creating configMap with name cm-test-opt-create-b6d0b565-7ec4-4d86-acaa-f49f6eba659d 02/25/23 11:23:26.665 + STEP: waiting to observe update in volume 02/25/23 11:23:26.672 + [AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 - Dec 14 16:55:21.747: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "resourcequota-9821" for this suite. 12/14/22 16:55:21.764 + Feb 25 11:24:53.557: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-391" for this suite. 02/25/23 11:24:53.576 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSS ------------------------------ -[sig-node] Secrets - should be consumable from pods in env vars [NodeConformance] [Conformance] - test/e2e/common/node/secrets.go:45 -[BeforeEach] [sig-node] Secrets +[sig-network] IngressClass API + should support creating IngressClass API operations [Conformance] + test/e2e/network/ingressclass.go:223 +[BeforeEach] [sig-network] IngressClass API test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:55:21.822 -Dec 14 16:55:21.823: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename secrets 12/14/22 16:55:21.824 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:21.861 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:21.869 -[It] should be consumable from pods in env vars [NodeConformance] [Conformance] - test/e2e/common/node/secrets.go:45 -STEP: Creating secret with name secret-test-1b1414df-b5d3-4f83-b05b-978f69d27a29 12/14/22 16:55:21.877 -STEP: Creating a pod to test consume secrets 12/14/22 16:55:21.885 -Dec 14 16:55:21.895: INFO: Waiting up to 5m0s for pod "pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7" in namespace "secrets-2754" to be "Succeeded or Failed" -Dec 14 16:55:21.900: INFO: Pod "pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7": Phase="Pending", Reason="", readiness=false. Elapsed: 4.792776ms -Dec 14 16:55:23.907: INFO: Pod "pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.0116574s -Dec 14 16:55:25.910: INFO: Pod "pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7": Phase="Pending", Reason="", readiness=false. Elapsed: 4.014677816s -Dec 14 16:55:27.907: INFO: Pod "pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.011617658s -STEP: Saw pod success 12/14/22 16:55:27.907 -Dec 14 16:55:27.907: INFO: Pod "pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7" satisfied condition "Succeeded or Failed" -Dec 14 16:55:27.910: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7 container secret-env-test: -STEP: delete the pod 12/14/22 16:55:28.285 -Dec 14 16:55:28.303: INFO: Waiting for pod pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7 to disappear -Dec 14 16:55:28.307: INFO: Pod pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7 no longer exists -[AfterEach] [sig-node] Secrets +STEP: Creating a kubernetes client 02/25/23 11:24:53.607 +Feb 25 11:24:53.607: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename ingressclass 02/25/23 11:24:53.612 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:24:53.641 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:24:53.645 +[BeforeEach] [sig-network] IngressClass API + test/e2e/network/ingressclass.go:211 +[It] should support creating IngressClass API operations [Conformance] + test/e2e/network/ingressclass.go:223 +STEP: getting /apis 02/25/23 11:24:53.65 +STEP: getting /apis/networking.k8s.io 02/25/23 11:24:53.654 +STEP: getting /apis/networking.k8s.iov1 02/25/23 11:24:53.656 +STEP: creating 02/25/23 11:24:53.658 +STEP: getting 02/25/23 11:24:53.689 +STEP: listing 02/25/23 11:24:53.696 +STEP: watching 02/25/23 11:24:53.701 +Feb 25 11:24:53.702: INFO: starting watch +STEP: patching 02/25/23 11:24:53.703 +STEP: updating 02/25/23 11:24:53.714 +Feb 25 11:24:53.722: INFO: waiting for watch events with expected annotations +Feb 25 11:24:53.722: INFO: saw patched and updated annotations +STEP: deleting 02/25/23 11:24:53.722 +STEP: deleting a collection 02/25/23 11:24:53.744 +[AfterEach] [sig-network] IngressClass API test/e2e/framework/framework.go:187 -Dec 14 16:55:28.308: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-2754" for this suite. 12/14/22 16:55:28.313 -{"msg":"PASSED [sig-node] Secrets should be consumable from pods in env vars [NodeConformance] [Conformance]","completed":286,"skipped":5508,"failed":0} +Feb 25 11:24:53.770: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "ingressclass-9391" for this suite. 02/25/23 11:24:53.777 +{"msg":"PASSED [sig-network] IngressClass API should support creating IngressClass API operations [Conformance]","completed":283,"skipped":5332,"failed":0} ------------------------------ -• [SLOW TEST] [6.500 seconds] -[sig-node] Secrets -test/e2e/common/node/framework.go:23 - should be consumable from pods in env vars [NodeConformance] [Conformance] - test/e2e/common/node/secrets.go:45 +• [0.182 seconds] +[sig-network] IngressClass API +test/e2e/network/common/framework.go:23 + should support creating IngressClass API operations [Conformance] + test/e2e/network/ingressclass.go:223 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Secrets + [BeforeEach] [sig-network] IngressClass API test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:55:21.822 - Dec 14 16:55:21.823: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename secrets 12/14/22 16:55:21.824 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:21.861 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:21.869 - [It] should be consumable from pods in env vars [NodeConformance] [Conformance] - test/e2e/common/node/secrets.go:45 - STEP: Creating secret with name secret-test-1b1414df-b5d3-4f83-b05b-978f69d27a29 12/14/22 16:55:21.877 - STEP: Creating a pod to test consume secrets 12/14/22 16:55:21.885 - Dec 14 16:55:21.895: INFO: Waiting up to 5m0s for pod "pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7" in namespace "secrets-2754" to be "Succeeded or Failed" - Dec 14 16:55:21.900: INFO: Pod "pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7": Phase="Pending", Reason="", readiness=false. Elapsed: 4.792776ms - Dec 14 16:55:23.907: INFO: Pod "pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.0116574s - Dec 14 16:55:25.910: INFO: Pod "pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7": Phase="Pending", Reason="", readiness=false. Elapsed: 4.014677816s - Dec 14 16:55:27.907: INFO: Pod "pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.011617658s - STEP: Saw pod success 12/14/22 16:55:27.907 - Dec 14 16:55:27.907: INFO: Pod "pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7" satisfied condition "Succeeded or Failed" - Dec 14 16:55:27.910: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7 container secret-env-test: - STEP: delete the pod 12/14/22 16:55:28.285 - Dec 14 16:55:28.303: INFO: Waiting for pod pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7 to disappear - Dec 14 16:55:28.307: INFO: Pod pod-secrets-1415a8a2-cbc0-46f8-bda2-b32e0c9195f7 no longer exists - [AfterEach] [sig-node] Secrets + STEP: Creating a kubernetes client 02/25/23 11:24:53.607 + Feb 25 11:24:53.607: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename ingressclass 02/25/23 11:24:53.612 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:24:53.641 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:24:53.645 + [BeforeEach] [sig-network] IngressClass API + test/e2e/network/ingressclass.go:211 + [It] should support creating IngressClass API operations [Conformance] + test/e2e/network/ingressclass.go:223 + STEP: getting /apis 02/25/23 11:24:53.65 + STEP: getting /apis/networking.k8s.io 02/25/23 11:24:53.654 + STEP: getting /apis/networking.k8s.iov1 02/25/23 11:24:53.656 + STEP: creating 02/25/23 11:24:53.658 + STEP: getting 02/25/23 11:24:53.689 + STEP: listing 02/25/23 11:24:53.696 + STEP: watching 02/25/23 11:24:53.701 + Feb 25 11:24:53.702: INFO: starting watch + STEP: patching 02/25/23 11:24:53.703 + STEP: updating 02/25/23 11:24:53.714 + Feb 25 11:24:53.722: INFO: waiting for watch events with expected annotations + Feb 25 11:24:53.722: INFO: saw patched and updated annotations + STEP: deleting 02/25/23 11:24:53.722 + STEP: deleting a collection 02/25/23 11:24:53.744 + [AfterEach] [sig-network] IngressClass API test/e2e/framework/framework.go:187 - Dec 14 16:55:28.308: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "secrets-2754" for this suite. 12/14/22 16:55:28.313 + Feb 25 11:24:53.770: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "ingressclass-9391" for this suite. 02/25/23 11:24:53.777 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Projected configMap - should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:374 -[BeforeEach] [sig-storage] Projected configMap - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:55:28.324 -Dec 14 16:55:28.324: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 16:55:28.329 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:28.355 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:28.358 -[It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:374 -STEP: Creating configMap with name projected-configmap-test-volume-9a8b3ff6-a2fb-4336-b0cb-854b899842d0 12/14/22 16:55:28.361 -STEP: Creating a pod to test consume configMaps 12/14/22 16:55:28.366 -Dec 14 16:55:28.387: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f" in namespace "projected-8009" to be "Succeeded or Failed" -Dec 14 16:55:28.391: INFO: Pod "pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.110466ms -Dec 14 16:55:30.398: INFO: Pod "pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.0112112s -Dec 14 16:55:32.397: INFO: Pod "pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.010178376s -Dec 14 16:55:34.399: INFO: Pod "pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.012145805s -STEP: Saw pod success 12/14/22 16:55:34.399 -Dec 14 16:55:34.399: INFO: Pod "pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f" satisfied condition "Succeeded or Failed" -Dec 14 16:55:34.405: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f container projected-configmap-volume-test: -STEP: delete the pod 12/14/22 16:55:34.423 -Dec 14 16:55:34.443: INFO: Waiting for pod pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f to disappear -Dec 14 16:55:34.447: INFO: Pod pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f no longer exists -[AfterEach] [sig-storage] Projected configMap +[sig-storage] CSIStorageCapacity + should support CSIStorageCapacities API operations [Conformance] + test/e2e/storage/csistoragecapacity.go:49 +[BeforeEach] [sig-storage] CSIStorageCapacity + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 11:24:53.798 +Feb 25 11:24:53.798: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename csistoragecapacity 02/25/23 11:24:53.801 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:24:53.829 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:24:53.835 +[It] should support CSIStorageCapacities API operations [Conformance] + test/e2e/storage/csistoragecapacity.go:49 +STEP: getting /apis 02/25/23 11:24:53.84 +STEP: getting /apis/storage.k8s.io 02/25/23 11:24:53.845 +STEP: getting /apis/storage.k8s.io/v1 02/25/23 11:24:53.848 +STEP: creating 02/25/23 11:24:53.85 +STEP: watching 02/25/23 11:24:53.888 +Feb 25 11:24:53.888: INFO: starting watch +STEP: getting 02/25/23 11:24:53.901 +STEP: listing in namespace 02/25/23 11:24:53.906 +STEP: listing across namespaces 02/25/23 11:24:53.912 +STEP: patching 02/25/23 11:24:53.92 +STEP: updating 02/25/23 11:24:53.932 +Feb 25 11:24:53.939: INFO: waiting for watch events with expected annotations in namespace +Feb 25 11:24:53.939: INFO: waiting for watch events with expected annotations across namespace +STEP: deleting 02/25/23 11:24:53.939 +STEP: deleting a collection 02/25/23 11:24:53.959 +[AfterEach] [sig-storage] CSIStorageCapacity test/e2e/framework/framework.go:187 -Dec 14 16:55:34.448: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-8009" for this suite. 12/14/22 16:55:34.455 -{"msg":"PASSED [sig-storage] Projected configMap should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance]","completed":287,"skipped":5523,"failed":0} +Feb 25 11:24:53.980: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "csistoragecapacity-9710" for this suite. 02/25/23 11:24:53.986 +{"msg":"PASSED [sig-storage] CSIStorageCapacity should support CSIStorageCapacities API operations [Conformance]","completed":284,"skipped":5391,"failed":0} ------------------------------ -• [SLOW TEST] [6.139 seconds] -[sig-storage] Projected configMap -test/e2e/common/storage/framework.go:23 - should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:374 +• [0.198 seconds] +[sig-storage] CSIStorageCapacity +test/e2e/storage/utils/framework.go:23 + should support CSIStorageCapacities API operations [Conformance] + test/e2e/storage/csistoragecapacity.go:49 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected configMap + [BeforeEach] [sig-storage] CSIStorageCapacity test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:55:28.324 - Dec 14 16:55:28.324: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 16:55:28.329 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:28.355 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:28.358 - [It] should be consumable in multiple volumes in the same pod [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:374 - STEP: Creating configMap with name projected-configmap-test-volume-9a8b3ff6-a2fb-4336-b0cb-854b899842d0 12/14/22 16:55:28.361 - STEP: Creating a pod to test consume configMaps 12/14/22 16:55:28.366 - Dec 14 16:55:28.387: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f" in namespace "projected-8009" to be "Succeeded or Failed" - Dec 14 16:55:28.391: INFO: Pod "pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.110466ms - Dec 14 16:55:30.398: INFO: Pod "pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.0112112s - Dec 14 16:55:32.397: INFO: Pod "pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f": Phase="Pending", Reason="", readiness=false. Elapsed: 4.010178376s - Dec 14 16:55:34.399: INFO: Pod "pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.012145805s - STEP: Saw pod success 12/14/22 16:55:34.399 - Dec 14 16:55:34.399: INFO: Pod "pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f" satisfied condition "Succeeded or Failed" - Dec 14 16:55:34.405: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f container projected-configmap-volume-test: - STEP: delete the pod 12/14/22 16:55:34.423 - Dec 14 16:55:34.443: INFO: Waiting for pod pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f to disappear - Dec 14 16:55:34.447: INFO: Pod pod-projected-configmaps-2c8f8639-1cd6-43d4-a07b-5b7b54ae615f no longer exists - [AfterEach] [sig-storage] Projected configMap + STEP: Creating a kubernetes client 02/25/23 11:24:53.798 + Feb 25 11:24:53.798: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename csistoragecapacity 02/25/23 11:24:53.801 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:24:53.829 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:24:53.835 + [It] should support CSIStorageCapacities API operations [Conformance] + test/e2e/storage/csistoragecapacity.go:49 + STEP: getting /apis 02/25/23 11:24:53.84 + STEP: getting /apis/storage.k8s.io 02/25/23 11:24:53.845 + STEP: getting /apis/storage.k8s.io/v1 02/25/23 11:24:53.848 + STEP: creating 02/25/23 11:24:53.85 + STEP: watching 02/25/23 11:24:53.888 + Feb 25 11:24:53.888: INFO: starting watch + STEP: getting 02/25/23 11:24:53.901 + STEP: listing in namespace 02/25/23 11:24:53.906 + STEP: listing across namespaces 02/25/23 11:24:53.912 + STEP: patching 02/25/23 11:24:53.92 + STEP: updating 02/25/23 11:24:53.932 + Feb 25 11:24:53.939: INFO: waiting for watch events with expected annotations in namespace + Feb 25 11:24:53.939: INFO: waiting for watch events with expected annotations across namespace + STEP: deleting 02/25/23 11:24:53.939 + STEP: deleting a collection 02/25/23 11:24:53.959 + [AfterEach] [sig-storage] CSIStorageCapacity test/e2e/framework/framework.go:187 - Dec 14 16:55:34.448: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-8009" for this suite. 12/14/22 16:55:34.455 + Feb 25 11:24:53.980: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "csistoragecapacity-9710" for this suite. 02/25/23 11:24:53.986 << End Captured GinkgoWriter Output ------------------------------ -S +SSSSSSSSSSSSSS ------------------------------ [sig-network] Services - should serve multiport endpoints from pods [Conformance] - test/e2e/network/service.go:852 + should be able to create a functioning NodePort service [Conformance] + test/e2e/network/service.go:1268 [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:55:34.465 -Dec 14 16:55:34.466: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 16:55:34.469 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:34.493 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:34.497 +STEP: Creating a kubernetes client 02/25/23 11:24:53.997 +Feb 25 11:24:53.998: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 11:24:54 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:24:54.02 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:24:54.023 [BeforeEach] [sig-network] Services test/e2e/network/service.go:758 -[It] should serve multiport endpoints from pods [Conformance] - test/e2e/network/service.go:852 -STEP: creating service multi-endpoint-test in namespace services-4338 12/14/22 16:55:34.5 -STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-4338 to expose endpoints map[] 12/14/22 16:55:34.514 -Dec 14 16:55:34.519: INFO: Failed go get Endpoints object: endpoints "multi-endpoint-test" not found -Dec 14 16:55:35.531: INFO: successfully validated that service multi-endpoint-test in namespace services-4338 exposes endpoints map[] -STEP: Creating pod pod1 in namespace services-4338 12/14/22 16:55:35.531 -Dec 14 16:55:35.540: INFO: Waiting up to 5m0s for pod "pod1" in namespace "services-4338" to be "running and ready" -Dec 14 16:55:35.543: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 3.404862ms -Dec 14 16:55:35.544: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:55:37.552: INFO: Pod "pod1": Phase="Running", Reason="", readiness=true. Elapsed: 2.012394674s -Dec 14 16:55:37.552: INFO: The phase of Pod pod1 is Running (Ready = true) -Dec 14 16:55:37.553: INFO: Pod "pod1" satisfied condition "running and ready" -STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-4338 to expose endpoints map[pod1:[100]] 12/14/22 16:55:37.558 -Dec 14 16:55:37.578: INFO: successfully validated that service multi-endpoint-test in namespace services-4338 exposes endpoints map[pod1:[100]] -STEP: Creating pod pod2 in namespace services-4338 12/14/22 16:55:37.578 -Dec 14 16:55:37.584: INFO: Waiting up to 5m0s for pod "pod2" in namespace "services-4338" to be "running and ready" -Dec 14 16:55:37.594: INFO: Pod "pod2": Phase="Pending", Reason="", readiness=false. Elapsed: 9.478467ms -Dec 14 16:55:37.594: INFO: The phase of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:55:39.600: INFO: Pod "pod2": Phase="Running", Reason="", readiness=true. Elapsed: 2.015556608s -Dec 14 16:55:39.600: INFO: The phase of Pod pod2 is Running (Ready = true) -Dec 14 16:55:39.600: INFO: Pod "pod2" satisfied condition "running and ready" -STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-4338 to expose endpoints map[pod1:[100] pod2:[101]] 12/14/22 16:55:39.624 -Dec 14 16:55:39.638: INFO: successfully validated that service multi-endpoint-test in namespace services-4338 exposes endpoints map[pod1:[100] pod2:[101]] -STEP: Checking if the Service forwards traffic to pods 12/14/22 16:55:39.638 -Dec 14 16:55:39.638: INFO: Creating new exec pod -Dec 14 16:55:39.646: INFO: Waiting up to 5m0s for pod "execpodthbk9" in namespace "services-4338" to be "running" -Dec 14 16:55:39.649: INFO: Pod "execpodthbk9": Phase="Pending", Reason="", readiness=false. Elapsed: 3.547782ms -Dec 14 16:55:41.659: INFO: Pod "execpodthbk9": Phase="Running", Reason="", readiness=true. Elapsed: 2.013097477s -Dec 14 16:55:41.659: INFO: Pod "execpodthbk9" satisfied condition "running" -Dec 14 16:55:42.661: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4338 exec execpodthbk9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 multi-endpoint-test 80' -Dec 14 16:55:42.939: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 multi-endpoint-test 80\nConnection to multi-endpoint-test 80 port [tcp/http] succeeded!\n" -Dec 14 16:55:42.939: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:55:42.939: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4338 exec execpodthbk9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.21.138 80' -Dec 14 16:55:43.229: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.21.138 80\nConnection to 10.233.21.138 80 port [tcp/http] succeeded!\n" -Dec 14 16:55:43.229: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:55:43.229: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4338 exec execpodthbk9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 multi-endpoint-test 81' -Dec 14 16:55:43.435: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 multi-endpoint-test 81\nConnection to multi-endpoint-test 81 port [tcp/*] succeeded!\n" -Dec 14 16:55:43.435: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -Dec 14 16:55:43.436: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4338 exec execpodthbk9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.21.138 81' -Dec 14 16:55:43.644: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.21.138 81\nConnection to 10.233.21.138 81 port [tcp/*] succeeded!\n" -Dec 14 16:55:43.644: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" -STEP: Deleting pod pod1 in namespace services-4338 12/14/22 16:55:43.645 -STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-4338 to expose endpoints map[pod2:[101]] 12/14/22 16:55:43.699 -Dec 14 16:55:44.739: INFO: successfully validated that service multi-endpoint-test in namespace services-4338 exposes endpoints map[pod2:[101]] -STEP: Deleting pod pod2 in namespace services-4338 12/14/22 16:55:44.739 -STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-4338 to expose endpoints map[] 12/14/22 16:55:44.759 -Dec 14 16:55:45.799: INFO: successfully validated that service multi-endpoint-test in namespace services-4338 exposes endpoints map[] +[It] should be able to create a functioning NodePort service [Conformance] + test/e2e/network/service.go:1268 +STEP: creating service nodeport-test with type=NodePort in namespace services-2322 02/25/23 11:24:54.027 +STEP: creating replication controller nodeport-test in namespace services-2322 02/25/23 11:24:54.047 +I0225 11:24:54.058170 13 runners.go:193] Created replication controller with name: nodeport-test, namespace: services-2322, replica count: 2 +I0225 11:24:57.111451 13 runners.go:193] nodeport-test Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +Feb 25 11:24:57.111: INFO: Creating new exec pod +Feb 25 11:24:57.130: INFO: Waiting up to 5m0s for pod "execpodz75tn" in namespace "services-2322" to be "running" +Feb 25 11:24:57.140: INFO: Pod "execpodz75tn": Phase="Pending", Reason="", readiness=false. Elapsed: 9.642442ms +Feb 25 11:24:59.158: INFO: Pod "execpodz75tn": Phase="Running", Reason="", readiness=true. Elapsed: 2.027918483s +Feb 25 11:24:59.158: INFO: Pod "execpodz75tn" satisfied condition "running" +Feb 25 11:25:00.186: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2322 exec execpodz75tn -- /bin/sh -x -c echo hostName | nc -v -t -w 2 nodeport-test 80' +Feb 25 11:25:00.553: INFO: stderr: "+ + nc -v -t -w 2 nodeport-test 80\necho hostName\nConnection to nodeport-test 80 port [tcp/http] succeeded!\n" +Feb 25 11:25:00.553: INFO: stdout: "nodeport-test-ngnq4" +Feb 25 11:25:00.559: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2322 exec execpodz75tn -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.47.204 80' +Feb 25 11:25:00.859: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.47.204 80\nConnection to 10.233.47.204 80 port [tcp/http] succeeded!\n" +Feb 25 11:25:00.859: INFO: stdout: "nodeport-test-dssd7" +Feb 25 11:25:00.860: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2322 exec execpodz75tn -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.44 32199' +Feb 25 11:25:01.506: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.44 32199\nConnection to 192.168.121.44 32199 port [tcp/*] succeeded!\n" +Feb 25 11:25:01.506: INFO: stdout: "nodeport-test-ngnq4" +Feb 25 11:25:01.506: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2322 exec execpodz75tn -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.250 32199' +Feb 25 11:25:01.776: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.250 32199\nConnection to 192.168.121.250 32199 port [tcp/*] succeeded!\n" +Feb 25 11:25:01.776: INFO: stdout: "nodeport-test-dssd7" [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 16:55:45.863: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-4338" for this suite. 12/14/22 16:55:45.87 +Feb 25 11:25:01.776: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-2322" for this suite. 02/25/23 11:25:01.783 [AfterEach] [sig-network] Services test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should serve multiport endpoints from pods [Conformance]","completed":288,"skipped":5524,"failed":0} +{"msg":"PASSED [sig-network] Services should be able to create a functioning NodePort service [Conformance]","completed":285,"skipped":5405,"failed":0} ------------------------------ -• [SLOW TEST] [11.416 seconds] +• [SLOW TEST] [7.794 seconds] [sig-network] Services test/e2e/network/common/framework.go:23 - should serve multiport endpoints from pods [Conformance] - test/e2e/network/service.go:852 + should be able to create a functioning NodePort service [Conformance] + test/e2e/network/service.go:1268 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:55:34.465 - Dec 14 16:55:34.466: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 16:55:34.469 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:34.493 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:34.497 + STEP: Creating a kubernetes client 02/25/23 11:24:53.997 + Feb 25 11:24:53.998: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 11:24:54 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:24:54.02 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:24:54.023 [BeforeEach] [sig-network] Services test/e2e/network/service.go:758 - [It] should serve multiport endpoints from pods [Conformance] - test/e2e/network/service.go:852 - STEP: creating service multi-endpoint-test in namespace services-4338 12/14/22 16:55:34.5 - STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-4338 to expose endpoints map[] 12/14/22 16:55:34.514 - Dec 14 16:55:34.519: INFO: Failed go get Endpoints object: endpoints "multi-endpoint-test" not found - Dec 14 16:55:35.531: INFO: successfully validated that service multi-endpoint-test in namespace services-4338 exposes endpoints map[] - STEP: Creating pod pod1 in namespace services-4338 12/14/22 16:55:35.531 - Dec 14 16:55:35.540: INFO: Waiting up to 5m0s for pod "pod1" in namespace "services-4338" to be "running and ready" - Dec 14 16:55:35.543: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 3.404862ms - Dec 14 16:55:35.544: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:55:37.552: INFO: Pod "pod1": Phase="Running", Reason="", readiness=true. Elapsed: 2.012394674s - Dec 14 16:55:37.552: INFO: The phase of Pod pod1 is Running (Ready = true) - Dec 14 16:55:37.553: INFO: Pod "pod1" satisfied condition "running and ready" - STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-4338 to expose endpoints map[pod1:[100]] 12/14/22 16:55:37.558 - Dec 14 16:55:37.578: INFO: successfully validated that service multi-endpoint-test in namespace services-4338 exposes endpoints map[pod1:[100]] - STEP: Creating pod pod2 in namespace services-4338 12/14/22 16:55:37.578 - Dec 14 16:55:37.584: INFO: Waiting up to 5m0s for pod "pod2" in namespace "services-4338" to be "running and ready" - Dec 14 16:55:37.594: INFO: Pod "pod2": Phase="Pending", Reason="", readiness=false. Elapsed: 9.478467ms - Dec 14 16:55:37.594: INFO: The phase of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:55:39.600: INFO: Pod "pod2": Phase="Running", Reason="", readiness=true. Elapsed: 2.015556608s - Dec 14 16:55:39.600: INFO: The phase of Pod pod2 is Running (Ready = true) - Dec 14 16:55:39.600: INFO: Pod "pod2" satisfied condition "running and ready" - STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-4338 to expose endpoints map[pod1:[100] pod2:[101]] 12/14/22 16:55:39.624 - Dec 14 16:55:39.638: INFO: successfully validated that service multi-endpoint-test in namespace services-4338 exposes endpoints map[pod1:[100] pod2:[101]] - STEP: Checking if the Service forwards traffic to pods 12/14/22 16:55:39.638 - Dec 14 16:55:39.638: INFO: Creating new exec pod - Dec 14 16:55:39.646: INFO: Waiting up to 5m0s for pod "execpodthbk9" in namespace "services-4338" to be "running" - Dec 14 16:55:39.649: INFO: Pod "execpodthbk9": Phase="Pending", Reason="", readiness=false. Elapsed: 3.547782ms - Dec 14 16:55:41.659: INFO: Pod "execpodthbk9": Phase="Running", Reason="", readiness=true. Elapsed: 2.013097477s - Dec 14 16:55:41.659: INFO: Pod "execpodthbk9" satisfied condition "running" - Dec 14 16:55:42.661: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4338 exec execpodthbk9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 multi-endpoint-test 80' - Dec 14 16:55:42.939: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 multi-endpoint-test 80\nConnection to multi-endpoint-test 80 port [tcp/http] succeeded!\n" - Dec 14 16:55:42.939: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:55:42.939: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4338 exec execpodthbk9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.21.138 80' - Dec 14 16:55:43.229: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.21.138 80\nConnection to 10.233.21.138 80 port [tcp/http] succeeded!\n" - Dec 14 16:55:43.229: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:55:43.229: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4338 exec execpodthbk9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 multi-endpoint-test 81' - Dec 14 16:55:43.435: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 multi-endpoint-test 81\nConnection to multi-endpoint-test 81 port [tcp/*] succeeded!\n" - Dec 14 16:55:43.435: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - Dec 14 16:55:43.436: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-4338 exec execpodthbk9 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.21.138 81' - Dec 14 16:55:43.644: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.21.138 81\nConnection to 10.233.21.138 81 port [tcp/*] succeeded!\n" - Dec 14 16:55:43.644: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" - STEP: Deleting pod pod1 in namespace services-4338 12/14/22 16:55:43.645 - STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-4338 to expose endpoints map[pod2:[101]] 12/14/22 16:55:43.699 - Dec 14 16:55:44.739: INFO: successfully validated that service multi-endpoint-test in namespace services-4338 exposes endpoints map[pod2:[101]] - STEP: Deleting pod pod2 in namespace services-4338 12/14/22 16:55:44.739 - STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-4338 to expose endpoints map[] 12/14/22 16:55:44.759 - Dec 14 16:55:45.799: INFO: successfully validated that service multi-endpoint-test in namespace services-4338 exposes endpoints map[] + [It] should be able to create a functioning NodePort service [Conformance] + test/e2e/network/service.go:1268 + STEP: creating service nodeport-test with type=NodePort in namespace services-2322 02/25/23 11:24:54.027 + STEP: creating replication controller nodeport-test in namespace services-2322 02/25/23 11:24:54.047 + I0225 11:24:54.058170 13 runners.go:193] Created replication controller with name: nodeport-test, namespace: services-2322, replica count: 2 + I0225 11:24:57.111451 13 runners.go:193] nodeport-test Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + Feb 25 11:24:57.111: INFO: Creating new exec pod + Feb 25 11:24:57.130: INFO: Waiting up to 5m0s for pod "execpodz75tn" in namespace "services-2322" to be "running" + Feb 25 11:24:57.140: INFO: Pod "execpodz75tn": Phase="Pending", Reason="", readiness=false. Elapsed: 9.642442ms + Feb 25 11:24:59.158: INFO: Pod "execpodz75tn": Phase="Running", Reason="", readiness=true. Elapsed: 2.027918483s + Feb 25 11:24:59.158: INFO: Pod "execpodz75tn" satisfied condition "running" + Feb 25 11:25:00.186: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2322 exec execpodz75tn -- /bin/sh -x -c echo hostName | nc -v -t -w 2 nodeport-test 80' + Feb 25 11:25:00.553: INFO: stderr: "+ + nc -v -t -w 2 nodeport-test 80\necho hostName\nConnection to nodeport-test 80 port [tcp/http] succeeded!\n" + Feb 25 11:25:00.553: INFO: stdout: "nodeport-test-ngnq4" + Feb 25 11:25:00.559: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2322 exec execpodz75tn -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.47.204 80' + Feb 25 11:25:00.859: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.47.204 80\nConnection to 10.233.47.204 80 port [tcp/http] succeeded!\n" + Feb 25 11:25:00.859: INFO: stdout: "nodeport-test-dssd7" + Feb 25 11:25:00.860: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2322 exec execpodz75tn -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.44 32199' + Feb 25 11:25:01.506: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.44 32199\nConnection to 192.168.121.44 32199 port [tcp/*] succeeded!\n" + Feb 25 11:25:01.506: INFO: stdout: "nodeport-test-ngnq4" + Feb 25 11:25:01.506: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-2322 exec execpodz75tn -- /bin/sh -x -c echo hostName | nc -v -t -w 2 192.168.121.250 32199' + Feb 25 11:25:01.776: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 192.168.121.250 32199\nConnection to 192.168.121.250 32199 port [tcp/*] succeeded!\n" + Feb 25 11:25:01.776: INFO: stdout: "nodeport-test-dssd7" [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 16:55:45.863: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-4338" for this suite. 12/14/22 16:55:45.87 + Feb 25 11:25:01.776: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-2322" for this suite. 02/25/23 11:25:01.783 [AfterEach] [sig-network] Services test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ SSSSSS ------------------------------ -[sig-node] Pods - should patch a pod status [Conformance] - test/e2e/common/node/pods.go:1082 -[BeforeEach] [sig-node] Pods - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:55:45.883 -Dec 14 16:55:45.883: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename pods 12/14/22 16:55:45.886 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:45.911 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:45.915 -[BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 -[It] should patch a pod status [Conformance] - test/e2e/common/node/pods.go:1082 -STEP: Create a pod 12/14/22 16:55:45.918 -Dec 14 16:55:45.930: INFO: Waiting up to 5m0s for pod "pod-bmllb" in namespace "pods-3458" to be "running" -Dec 14 16:55:45.936: INFO: Pod "pod-bmllb": Phase="Pending", Reason="", readiness=false. Elapsed: 6.49001ms -Dec 14 16:55:47.941: INFO: Pod "pod-bmllb": Phase="Running", Reason="", readiness=true. Elapsed: 2.011910946s -Dec 14 16:55:47.942: INFO: Pod "pod-bmllb" satisfied condition "running" -STEP: patching /status 12/14/22 16:55:47.942 -Dec 14 16:55:47.951: INFO: Status Message: "Patched by e2e test" and Reason: "E2E" -[AfterEach] [sig-node] Pods - test/e2e/framework/framework.go:187 -Dec 14 16:55:47.951: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-3458" for this suite. 12/14/22 16:55:47.956 -{"msg":"PASSED [sig-node] Pods should patch a pod status [Conformance]","completed":289,"skipped":5530,"failed":0} ------------------------------- -• [2.082 seconds] -[sig-node] Pods -test/e2e/common/node/framework.go:23 - should patch a pod status [Conformance] - test/e2e/common/node/pods.go:1082 - - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Pods - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:55:45.883 - Dec 14 16:55:45.883: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename pods 12/14/22 16:55:45.886 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:45.911 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:45.915 - [BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 - [It] should patch a pod status [Conformance] - test/e2e/common/node/pods.go:1082 - STEP: Create a pod 12/14/22 16:55:45.918 - Dec 14 16:55:45.930: INFO: Waiting up to 5m0s for pod "pod-bmllb" in namespace "pods-3458" to be "running" - Dec 14 16:55:45.936: INFO: Pod "pod-bmllb": Phase="Pending", Reason="", readiness=false. Elapsed: 6.49001ms - Dec 14 16:55:47.941: INFO: Pod "pod-bmllb": Phase="Running", Reason="", readiness=true. Elapsed: 2.011910946s - Dec 14 16:55:47.942: INFO: Pod "pod-bmllb" satisfied condition "running" - STEP: patching /status 12/14/22 16:55:47.942 - Dec 14 16:55:47.951: INFO: Status Message: "Patched by e2e test" and Reason: "E2E" - [AfterEach] [sig-node] Pods - test/e2e/framework/framework.go:187 - Dec 14 16:55:47.951: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "pods-3458" for this suite. 12/14/22 16:55:47.956 - << End Captured GinkgoWriter Output ------------------------------- -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-api-machinery] Servers with support for Table transformation - should return a 406 for a backend which does not implement metadata [Conformance] - test/e2e/apimachinery/table_conversion.go:154 -[BeforeEach] [sig-api-machinery] Servers with support for Table transformation +[sig-apps] CronJob + should schedule multiple jobs concurrently [Conformance] + test/e2e/apps/cronjob.go:69 +[BeforeEach] [sig-apps] CronJob test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:55:47.979 -Dec 14 16:55:47.980: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename tables 12/14/22 16:55:47.986 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:48.016 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:48.02 -[BeforeEach] [sig-api-machinery] Servers with support for Table transformation - test/e2e/apimachinery/table_conversion.go:49 -[It] should return a 406 for a backend which does not implement metadata [Conformance] - test/e2e/apimachinery/table_conversion.go:154 -[AfterEach] [sig-api-machinery] Servers with support for Table transformation +STEP: Creating a kubernetes client 02/25/23 11:25:01.793 +Feb 25 11:25:01.793: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename cronjob 02/25/23 11:25:01.794 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:25:01.813 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:25:01.817 +[It] should schedule multiple jobs concurrently [Conformance] + test/e2e/apps/cronjob.go:69 +STEP: Creating a cronjob 02/25/23 11:25:01.82 +STEP: Ensuring more than one job is running at a time 02/25/23 11:25:01.828 +STEP: Ensuring at least two running jobs exists by listing jobs explicitly 02/25/23 11:27:01.836 +STEP: Removing cronjob 02/25/23 11:27:01.841 +[AfterEach] [sig-apps] CronJob test/e2e/framework/framework.go:187 -Dec 14 16:55:48.028: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "tables-2536" for this suite. 12/14/22 16:55:48.032 -{"msg":"PASSED [sig-api-machinery] Servers with support for Table transformation should return a 406 for a backend which does not implement metadata [Conformance]","completed":290,"skipped":5578,"failed":0} +Feb 25 11:27:01.859: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "cronjob-6649" for this suite. 02/25/23 11:27:01.866 +{"msg":"PASSED [sig-apps] CronJob should schedule multiple jobs concurrently [Conformance]","completed":286,"skipped":5411,"failed":0} ------------------------------ -• [0.060 seconds] -[sig-api-machinery] Servers with support for Table transformation -test/e2e/apimachinery/framework.go:23 - should return a 406 for a backend which does not implement metadata [Conformance] - test/e2e/apimachinery/table_conversion.go:154 +• [SLOW TEST] [120.092 seconds] +[sig-apps] CronJob +test/e2e/apps/framework.go:23 + should schedule multiple jobs concurrently [Conformance] + test/e2e/apps/cronjob.go:69 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Servers with support for Table transformation + [BeforeEach] [sig-apps] CronJob test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:55:47.979 - Dec 14 16:55:47.980: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename tables 12/14/22 16:55:47.986 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:48.016 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:48.02 - [BeforeEach] [sig-api-machinery] Servers with support for Table transformation - test/e2e/apimachinery/table_conversion.go:49 - [It] should return a 406 for a backend which does not implement metadata [Conformance] - test/e2e/apimachinery/table_conversion.go:154 - [AfterEach] [sig-api-machinery] Servers with support for Table transformation + STEP: Creating a kubernetes client 02/25/23 11:25:01.793 + Feb 25 11:25:01.793: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename cronjob 02/25/23 11:25:01.794 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:25:01.813 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:25:01.817 + [It] should schedule multiple jobs concurrently [Conformance] + test/e2e/apps/cronjob.go:69 + STEP: Creating a cronjob 02/25/23 11:25:01.82 + STEP: Ensuring more than one job is running at a time 02/25/23 11:25:01.828 + STEP: Ensuring at least two running jobs exists by listing jobs explicitly 02/25/23 11:27:01.836 + STEP: Removing cronjob 02/25/23 11:27:01.841 + [AfterEach] [sig-apps] CronJob test/e2e/framework/framework.go:187 - Dec 14 16:55:48.028: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "tables-2536" for this suite. 12/14/22 16:55:48.032 + Feb 25 11:27:01.859: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "cronjob-6649" for this suite. 02/25/23 11:27:01.866 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSS ------------------------------ -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - should mutate configmap [Conformance] - test/e2e/apimachinery/webhook.go:251 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +[sig-cli] Kubectl client Kubectl logs + should be able to retrieve and filter logs [Conformance] + test/e2e/kubectl/kubectl.go:1590 +[BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:55:48.049 -Dec 14 16:55:48.050: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 16:55:48.054 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:48.087 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:48.091 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 16:55:48.12 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:55:48.861 -STEP: Deploying the webhook pod 12/14/22 16:55:48.874 -STEP: Wait for the deployment to be ready 12/14/22 16:55:48.892 -Dec 14 16:55:48.901: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created -STEP: Deploying the webhook service 12/14/22 16:55:50.919 -STEP: Verifying the service has paired with the endpoint 12/14/22 16:55:50.934 -Dec 14 16:55:51.934: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] should mutate configmap [Conformance] - test/e2e/apimachinery/webhook.go:251 -STEP: Registering the mutating configmap webhook via the AdmissionRegistration API 12/14/22 16:55:51.939 -STEP: create a configmap that should be updated by the webhook 12/14/22 16:55:51.96 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 11:27:01.888 +Feb 25 11:27:01.888: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 11:27:01.892 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:01.921 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:01.927 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 +[BeforeEach] Kubectl logs + test/e2e/kubectl/kubectl.go:1570 +STEP: creating an pod 02/25/23 11:27:01.933 +Feb 25 11:27:01.934: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 run logs-generator --image=registry.k8s.io/e2e-test-images/agnhost:2.40 --restart=Never --pod-running-timeout=2m0s -- logs-generator --log-lines-total 100 --run-duration 20s' +Feb 25 11:27:02.081: INFO: stderr: "" +Feb 25 11:27:02.082: INFO: stdout: "pod/logs-generator created\n" +[It] should be able to retrieve and filter logs [Conformance] + test/e2e/kubectl/kubectl.go:1590 +STEP: Waiting for log generator to start. 02/25/23 11:27:02.082 +Feb 25 11:27:02.082: INFO: Waiting up to 5m0s for 1 pods to be running and ready, or succeeded: [logs-generator] +Feb 25 11:27:02.082: INFO: Waiting up to 5m0s for pod "logs-generator" in namespace "kubectl-3752" to be "running and ready, or succeeded" +Feb 25 11:27:02.091: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 8.679751ms +Feb 25 11:27:02.091: INFO: Error evaluating pod condition running and ready, or succeeded: want pod 'logs-generator' on 'diek3dua9gay-2' to be 'Running' but was 'Pending' +Feb 25 11:27:04.096: INFO: Pod "logs-generator": Phase="Running", Reason="", readiness=true. Elapsed: 2.014351366s +Feb 25 11:27:04.096: INFO: Pod "logs-generator" satisfied condition "running and ready, or succeeded" +Feb 25 11:27:04.096: INFO: Wanted all 1 pods to be running and ready, or succeeded. Result: true. Pods: [logs-generator] +STEP: checking for a matching strings 02/25/23 11:27:04.096 +Feb 25 11:27:04.097: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 logs logs-generator logs-generator' +Feb 25 11:27:04.283: INFO: stderr: "" +Feb 25 11:27:04.283: INFO: stdout: "I0225 11:27:03.234071 1 logs_generator.go:76] 0 GET /api/v1/namespaces/ns/pods/msk 249\nI0225 11:27:03.434469 1 logs_generator.go:76] 1 GET /api/v1/namespaces/ns/pods/vtl 385\nI0225 11:27:03.635231 1 logs_generator.go:76] 2 PUT /api/v1/namespaces/default/pods/4hl 316\nI0225 11:27:03.834591 1 logs_generator.go:76] 3 POST /api/v1/namespaces/ns/pods/5k8 448\nI0225 11:27:04.034930 1 logs_generator.go:76] 4 POST /api/v1/namespaces/ns/pods/44b9 246\nI0225 11:27:04.234518 1 logs_generator.go:76] 5 POST /api/v1/namespaces/kube-system/pods/6cl 263\n" +STEP: limiting log lines 02/25/23 11:27:04.283 +Feb 25 11:27:04.283: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 logs logs-generator logs-generator --tail=1' +Feb 25 11:27:04.507: INFO: stderr: "" +Feb 25 11:27:04.507: INFO: stdout: "I0225 11:27:04.434930 1 logs_generator.go:76] 6 PUT /api/v1/namespaces/kube-system/pods/zjxm 300\n" +Feb 25 11:27:04.507: INFO: got output "I0225 11:27:04.434930 1 logs_generator.go:76] 6 PUT /api/v1/namespaces/kube-system/pods/zjxm 300\n" +STEP: limiting log bytes 02/25/23 11:27:04.507 +Feb 25 11:27:04.508: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 logs logs-generator logs-generator --limit-bytes=1' +Feb 25 11:27:04.753: INFO: stderr: "" +Feb 25 11:27:04.753: INFO: stdout: "I" +Feb 25 11:27:04.753: INFO: got output "I" +STEP: exposing timestamps 02/25/23 11:27:04.753 +Feb 25 11:27:04.754: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 logs logs-generator logs-generator --tail=1 --timestamps' +Feb 25 11:27:05.085: INFO: stderr: "" +Feb 25 11:27:05.085: INFO: stdout: "2023-02-25T11:27:05.039673329Z I0225 11:27:05.039479 1 logs_generator.go:76] 9 PUT /api/v1/namespaces/default/pods/lhhr 292\n" +Feb 25 11:27:05.085: INFO: got output "2023-02-25T11:27:05.039673329Z I0225 11:27:05.039479 1 logs_generator.go:76] 9 PUT /api/v1/namespaces/default/pods/lhhr 292\n" +STEP: restricting to a time range 02/25/23 11:27:05.085 +Feb 25 11:27:07.586: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 logs logs-generator logs-generator --since=1s' +Feb 25 11:27:07.775: INFO: stderr: "" +Feb 25 11:27:07.775: INFO: stdout: "I0225 11:27:06.834304 1 logs_generator.go:76] 18 GET /api/v1/namespaces/default/pods/98q 548\nI0225 11:27:07.034739 1 logs_generator.go:76] 19 GET /api/v1/namespaces/kube-system/pods/tb94 216\nI0225 11:27:07.235128 1 logs_generator.go:76] 20 GET /api/v1/namespaces/ns/pods/v49 282\nI0225 11:27:07.434495 1 logs_generator.go:76] 21 POST /api/v1/namespaces/default/pods/grzt 364\nI0225 11:27:07.634965 1 logs_generator.go:76] 22 PUT /api/v1/namespaces/kube-system/pods/bq2m 226\n" +Feb 25 11:27:07.776: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 logs logs-generator logs-generator --since=24h' +Feb 25 11:27:07.941: INFO: stderr: "" +Feb 25 11:27:07.941: INFO: stdout: "I0225 11:27:03.234071 1 logs_generator.go:76] 0 GET /api/v1/namespaces/ns/pods/msk 249\nI0225 11:27:03.434469 1 logs_generator.go:76] 1 GET /api/v1/namespaces/ns/pods/vtl 385\nI0225 11:27:03.635231 1 logs_generator.go:76] 2 PUT /api/v1/namespaces/default/pods/4hl 316\nI0225 11:27:03.834591 1 logs_generator.go:76] 3 POST /api/v1/namespaces/ns/pods/5k8 448\nI0225 11:27:04.034930 1 logs_generator.go:76] 4 POST /api/v1/namespaces/ns/pods/44b9 246\nI0225 11:27:04.234518 1 logs_generator.go:76] 5 POST /api/v1/namespaces/kube-system/pods/6cl 263\nI0225 11:27:04.434930 1 logs_generator.go:76] 6 PUT /api/v1/namespaces/kube-system/pods/zjxm 300\nI0225 11:27:04.634361 1 logs_generator.go:76] 7 GET /api/v1/namespaces/kube-system/pods/574j 476\nI0225 11:27:04.834869 1 logs_generator.go:76] 8 POST /api/v1/namespaces/default/pods/8db9 508\nI0225 11:27:05.039479 1 logs_generator.go:76] 9 PUT /api/v1/namespaces/default/pods/lhhr 292\nI0225 11:27:05.235161 1 logs_generator.go:76] 10 PUT /api/v1/namespaces/default/pods/qbs 564\nI0225 11:27:05.434854 1 logs_generator.go:76] 11 PUT /api/v1/namespaces/ns/pods/5m2g 287\nI0225 11:27:05.634252 1 logs_generator.go:76] 12 GET /api/v1/namespaces/kube-system/pods/87m 491\nI0225 11:27:05.834753 1 logs_generator.go:76] 13 PUT /api/v1/namespaces/default/pods/mrf 233\nI0225 11:27:06.034228 1 logs_generator.go:76] 14 PUT /api/v1/namespaces/default/pods/9pf 470\nI0225 11:27:06.234693 1 logs_generator.go:76] 15 PUT /api/v1/namespaces/kube-system/pods/9f6 418\nI0225 11:27:06.434241 1 logs_generator.go:76] 16 GET /api/v1/namespaces/ns/pods/7ckl 397\nI0225 11:27:06.634804 1 logs_generator.go:76] 17 GET /api/v1/namespaces/kube-system/pods/jxd2 261\nI0225 11:27:06.834304 1 logs_generator.go:76] 18 GET /api/v1/namespaces/default/pods/98q 548\nI0225 11:27:07.034739 1 logs_generator.go:76] 19 GET /api/v1/namespaces/kube-system/pods/tb94 216\nI0225 11:27:07.235128 1 logs_generator.go:76] 20 GET /api/v1/namespaces/ns/pods/v49 282\nI0225 11:27:07.434495 1 logs_generator.go:76] 21 POST /api/v1/namespaces/default/pods/grzt 364\nI0225 11:27:07.634965 1 logs_generator.go:76] 22 PUT /api/v1/namespaces/kube-system/pods/bq2m 226\nI0225 11:27:07.834267 1 logs_generator.go:76] 23 PUT /api/v1/namespaces/kube-system/pods/7bk 393\n" +[AfterEach] Kubectl logs + test/e2e/kubectl/kubectl.go:1575 +Feb 25 11:27:07.942: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 delete pod logs-generator' +Feb 25 11:27:08.524: INFO: stderr: "" +Feb 25 11:27:08.524: INFO: stdout: "pod \"logs-generator\" deleted\n" +[AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 -Dec 14 16:55:51.988: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-357" for this suite. 12/14/22 16:55:51.994 -STEP: Destroying namespace "webhook-357-markers" for this suite. 12/14/22 16:55:52.001 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate configmap [Conformance]","completed":291,"skipped":5605,"failed":0} +Feb 25 11:27:08.524: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-3752" for this suite. 02/25/23 11:27:08.532 +{"msg":"PASSED [sig-cli] Kubectl client Kubectl logs should be able to retrieve and filter logs [Conformance]","completed":287,"skipped":5416,"failed":0} ------------------------------ -• [4.153 seconds] -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - should mutate configmap [Conformance] - test/e2e/apimachinery/webhook.go:251 +• [SLOW TEST] [6.653 seconds] +[sig-cli] Kubectl client +test/e2e/kubectl/framework.go:23 + Kubectl logs + test/e2e/kubectl/kubectl.go:1567 + should be able to retrieve and filter logs [Conformance] + test/e2e/kubectl/kubectl.go:1590 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + [BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:55:48.049 - Dec 14 16:55:48.050: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 16:55:48.054 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:48.087 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:48.091 - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 16:55:48.12 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 16:55:48.861 - STEP: Deploying the webhook pod 12/14/22 16:55:48.874 - STEP: Wait for the deployment to be ready 12/14/22 16:55:48.892 - Dec 14 16:55:48.901: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created - STEP: Deploying the webhook service 12/14/22 16:55:50.919 - STEP: Verifying the service has paired with the endpoint 12/14/22 16:55:50.934 - Dec 14 16:55:51.934: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] should mutate configmap [Conformance] - test/e2e/apimachinery/webhook.go:251 - STEP: Registering the mutating configmap webhook via the AdmissionRegistration API 12/14/22 16:55:51.939 - STEP: create a configmap that should be updated by the webhook 12/14/22 16:55:51.96 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 11:27:01.888 + Feb 25 11:27:01.888: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 11:27:01.892 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:01.921 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:01.927 + [BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 + [BeforeEach] Kubectl logs + test/e2e/kubectl/kubectl.go:1570 + STEP: creating an pod 02/25/23 11:27:01.933 + Feb 25 11:27:01.934: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 run logs-generator --image=registry.k8s.io/e2e-test-images/agnhost:2.40 --restart=Never --pod-running-timeout=2m0s -- logs-generator --log-lines-total 100 --run-duration 20s' + Feb 25 11:27:02.081: INFO: stderr: "" + Feb 25 11:27:02.082: INFO: stdout: "pod/logs-generator created\n" + [It] should be able to retrieve and filter logs [Conformance] + test/e2e/kubectl/kubectl.go:1590 + STEP: Waiting for log generator to start. 02/25/23 11:27:02.082 + Feb 25 11:27:02.082: INFO: Waiting up to 5m0s for 1 pods to be running and ready, or succeeded: [logs-generator] + Feb 25 11:27:02.082: INFO: Waiting up to 5m0s for pod "logs-generator" in namespace "kubectl-3752" to be "running and ready, or succeeded" + Feb 25 11:27:02.091: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 8.679751ms + Feb 25 11:27:02.091: INFO: Error evaluating pod condition running and ready, or succeeded: want pod 'logs-generator' on 'diek3dua9gay-2' to be 'Running' but was 'Pending' + Feb 25 11:27:04.096: INFO: Pod "logs-generator": Phase="Running", Reason="", readiness=true. Elapsed: 2.014351366s + Feb 25 11:27:04.096: INFO: Pod "logs-generator" satisfied condition "running and ready, or succeeded" + Feb 25 11:27:04.096: INFO: Wanted all 1 pods to be running and ready, or succeeded. Result: true. Pods: [logs-generator] + STEP: checking for a matching strings 02/25/23 11:27:04.096 + Feb 25 11:27:04.097: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 logs logs-generator logs-generator' + Feb 25 11:27:04.283: INFO: stderr: "" + Feb 25 11:27:04.283: INFO: stdout: "I0225 11:27:03.234071 1 logs_generator.go:76] 0 GET /api/v1/namespaces/ns/pods/msk 249\nI0225 11:27:03.434469 1 logs_generator.go:76] 1 GET /api/v1/namespaces/ns/pods/vtl 385\nI0225 11:27:03.635231 1 logs_generator.go:76] 2 PUT /api/v1/namespaces/default/pods/4hl 316\nI0225 11:27:03.834591 1 logs_generator.go:76] 3 POST /api/v1/namespaces/ns/pods/5k8 448\nI0225 11:27:04.034930 1 logs_generator.go:76] 4 POST /api/v1/namespaces/ns/pods/44b9 246\nI0225 11:27:04.234518 1 logs_generator.go:76] 5 POST /api/v1/namespaces/kube-system/pods/6cl 263\n" + STEP: limiting log lines 02/25/23 11:27:04.283 + Feb 25 11:27:04.283: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 logs logs-generator logs-generator --tail=1' + Feb 25 11:27:04.507: INFO: stderr: "" + Feb 25 11:27:04.507: INFO: stdout: "I0225 11:27:04.434930 1 logs_generator.go:76] 6 PUT /api/v1/namespaces/kube-system/pods/zjxm 300\n" + Feb 25 11:27:04.507: INFO: got output "I0225 11:27:04.434930 1 logs_generator.go:76] 6 PUT /api/v1/namespaces/kube-system/pods/zjxm 300\n" + STEP: limiting log bytes 02/25/23 11:27:04.507 + Feb 25 11:27:04.508: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 logs logs-generator logs-generator --limit-bytes=1' + Feb 25 11:27:04.753: INFO: stderr: "" + Feb 25 11:27:04.753: INFO: stdout: "I" + Feb 25 11:27:04.753: INFO: got output "I" + STEP: exposing timestamps 02/25/23 11:27:04.753 + Feb 25 11:27:04.754: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 logs logs-generator logs-generator --tail=1 --timestamps' + Feb 25 11:27:05.085: INFO: stderr: "" + Feb 25 11:27:05.085: INFO: stdout: "2023-02-25T11:27:05.039673329Z I0225 11:27:05.039479 1 logs_generator.go:76] 9 PUT /api/v1/namespaces/default/pods/lhhr 292\n" + Feb 25 11:27:05.085: INFO: got output "2023-02-25T11:27:05.039673329Z I0225 11:27:05.039479 1 logs_generator.go:76] 9 PUT /api/v1/namespaces/default/pods/lhhr 292\n" + STEP: restricting to a time range 02/25/23 11:27:05.085 + Feb 25 11:27:07.586: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 logs logs-generator logs-generator --since=1s' + Feb 25 11:27:07.775: INFO: stderr: "" + Feb 25 11:27:07.775: INFO: stdout: "I0225 11:27:06.834304 1 logs_generator.go:76] 18 GET /api/v1/namespaces/default/pods/98q 548\nI0225 11:27:07.034739 1 logs_generator.go:76] 19 GET /api/v1/namespaces/kube-system/pods/tb94 216\nI0225 11:27:07.235128 1 logs_generator.go:76] 20 GET /api/v1/namespaces/ns/pods/v49 282\nI0225 11:27:07.434495 1 logs_generator.go:76] 21 POST /api/v1/namespaces/default/pods/grzt 364\nI0225 11:27:07.634965 1 logs_generator.go:76] 22 PUT /api/v1/namespaces/kube-system/pods/bq2m 226\n" + Feb 25 11:27:07.776: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 logs logs-generator logs-generator --since=24h' + Feb 25 11:27:07.941: INFO: stderr: "" + Feb 25 11:27:07.941: INFO: stdout: "I0225 11:27:03.234071 1 logs_generator.go:76] 0 GET /api/v1/namespaces/ns/pods/msk 249\nI0225 11:27:03.434469 1 logs_generator.go:76] 1 GET /api/v1/namespaces/ns/pods/vtl 385\nI0225 11:27:03.635231 1 logs_generator.go:76] 2 PUT /api/v1/namespaces/default/pods/4hl 316\nI0225 11:27:03.834591 1 logs_generator.go:76] 3 POST /api/v1/namespaces/ns/pods/5k8 448\nI0225 11:27:04.034930 1 logs_generator.go:76] 4 POST /api/v1/namespaces/ns/pods/44b9 246\nI0225 11:27:04.234518 1 logs_generator.go:76] 5 POST /api/v1/namespaces/kube-system/pods/6cl 263\nI0225 11:27:04.434930 1 logs_generator.go:76] 6 PUT /api/v1/namespaces/kube-system/pods/zjxm 300\nI0225 11:27:04.634361 1 logs_generator.go:76] 7 GET /api/v1/namespaces/kube-system/pods/574j 476\nI0225 11:27:04.834869 1 logs_generator.go:76] 8 POST /api/v1/namespaces/default/pods/8db9 508\nI0225 11:27:05.039479 1 logs_generator.go:76] 9 PUT /api/v1/namespaces/default/pods/lhhr 292\nI0225 11:27:05.235161 1 logs_generator.go:76] 10 PUT /api/v1/namespaces/default/pods/qbs 564\nI0225 11:27:05.434854 1 logs_generator.go:76] 11 PUT /api/v1/namespaces/ns/pods/5m2g 287\nI0225 11:27:05.634252 1 logs_generator.go:76] 12 GET /api/v1/namespaces/kube-system/pods/87m 491\nI0225 11:27:05.834753 1 logs_generator.go:76] 13 PUT /api/v1/namespaces/default/pods/mrf 233\nI0225 11:27:06.034228 1 logs_generator.go:76] 14 PUT /api/v1/namespaces/default/pods/9pf 470\nI0225 11:27:06.234693 1 logs_generator.go:76] 15 PUT /api/v1/namespaces/kube-system/pods/9f6 418\nI0225 11:27:06.434241 1 logs_generator.go:76] 16 GET /api/v1/namespaces/ns/pods/7ckl 397\nI0225 11:27:06.634804 1 logs_generator.go:76] 17 GET /api/v1/namespaces/kube-system/pods/jxd2 261\nI0225 11:27:06.834304 1 logs_generator.go:76] 18 GET /api/v1/namespaces/default/pods/98q 548\nI0225 11:27:07.034739 1 logs_generator.go:76] 19 GET /api/v1/namespaces/kube-system/pods/tb94 216\nI0225 11:27:07.235128 1 logs_generator.go:76] 20 GET /api/v1/namespaces/ns/pods/v49 282\nI0225 11:27:07.434495 1 logs_generator.go:76] 21 POST /api/v1/namespaces/default/pods/grzt 364\nI0225 11:27:07.634965 1 logs_generator.go:76] 22 PUT /api/v1/namespaces/kube-system/pods/bq2m 226\nI0225 11:27:07.834267 1 logs_generator.go:76] 23 PUT /api/v1/namespaces/kube-system/pods/7bk 393\n" + [AfterEach] Kubectl logs + test/e2e/kubectl/kubectl.go:1575 + Feb 25 11:27:07.942: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-3752 delete pod logs-generator' + Feb 25 11:27:08.524: INFO: stderr: "" + Feb 25 11:27:08.524: INFO: stdout: "pod \"logs-generator\" deleted\n" + [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 16:55:51.988: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-357" for this suite. 12/14/22 16:55:51.994 - STEP: Destroying namespace "webhook-357-markers" for this suite. 12/14/22 16:55:52.001 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 + Feb 25 11:27:08.524: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-3752" for this suite. 02/25/23 11:27:08.532 << End Captured GinkgoWriter Output ------------------------------ SSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:186 -[BeforeEach] [sig-storage] EmptyDir volumes +[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition + getting/updating/patching custom resource definition status sub-resource works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:145 +[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:55:52.205 -Dec 14 16:55:52.205: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir 12/14/22 16:55:52.235 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:52.292 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:52.296 -[It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:186 -STEP: Creating a pod to test emptydir 0777 on node default medium 12/14/22 16:55:52.317 -Dec 14 16:55:52.365: INFO: Waiting up to 5m0s for pod "pod-3c618720-44b1-4fbb-8561-6aca1193652a" in namespace "emptydir-3422" to be "Succeeded or Failed" -Dec 14 16:55:52.371: INFO: Pod "pod-3c618720-44b1-4fbb-8561-6aca1193652a": Phase="Pending", Reason="", readiness=false. Elapsed: 5.994968ms -Dec 14 16:55:54.399: INFO: Pod "pod-3c618720-44b1-4fbb-8561-6aca1193652a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034124044s -Dec 14 16:55:56.391: INFO: Pod "pod-3c618720-44b1-4fbb-8561-6aca1193652a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.025732343s -Dec 14 16:55:58.392: INFO: Pod "pod-3c618720-44b1-4fbb-8561-6aca1193652a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.027021707s -STEP: Saw pod success 12/14/22 16:55:58.392 -Dec 14 16:55:58.393: INFO: Pod "pod-3c618720-44b1-4fbb-8561-6aca1193652a" satisfied condition "Succeeded or Failed" -Dec 14 16:55:58.400: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-3c618720-44b1-4fbb-8561-6aca1193652a container test-container: -STEP: delete the pod 12/14/22 16:55:58.41 -Dec 14 16:55:58.427: INFO: Waiting for pod pod-3c618720-44b1-4fbb-8561-6aca1193652a to disappear -Dec 14 16:55:58.433: INFO: Pod pod-3c618720-44b1-4fbb-8561-6aca1193652a no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +STEP: Creating a kubernetes client 02/25/23 11:27:08.542 +Feb 25 11:27:08.542: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename custom-resource-definition 02/25/23 11:27:08.545 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:08.567 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:08.571 +[It] getting/updating/patching custom resource definition status sub-resource works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:145 +Feb 25 11:27:08.577: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:55:58.433: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-3422" for this suite. 12/14/22 16:55:58.439 -{"msg":"PASSED [sig-storage] EmptyDir volumes should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance]","completed":292,"skipped":5608,"failed":0} +Feb 25 11:27:09.354: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "custom-resource-definition-8174" for this suite. 02/25/23 11:27:09.399 +{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition getting/updating/patching custom resource definition status sub-resource works [Conformance]","completed":288,"skipped":5419,"failed":0} ------------------------------ -• [SLOW TEST] [6.243 seconds] -[sig-storage] EmptyDir volumes -test/e2e/common/storage/framework.go:23 - should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:186 +• [0.944 seconds] +[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + Simple CustomResourceDefinition + test/e2e/apimachinery/custom_resource_definition.go:50 + getting/updating/patching custom resource definition status sub-resource works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:145 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir volumes + [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:55:52.205 - Dec 14 16:55:52.205: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir 12/14/22 16:55:52.235 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:52.292 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:52.296 - [It] should support (root,0777,default) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:186 - STEP: Creating a pod to test emptydir 0777 on node default medium 12/14/22 16:55:52.317 - Dec 14 16:55:52.365: INFO: Waiting up to 5m0s for pod "pod-3c618720-44b1-4fbb-8561-6aca1193652a" in namespace "emptydir-3422" to be "Succeeded or Failed" - Dec 14 16:55:52.371: INFO: Pod "pod-3c618720-44b1-4fbb-8561-6aca1193652a": Phase="Pending", Reason="", readiness=false. Elapsed: 5.994968ms - Dec 14 16:55:54.399: INFO: Pod "pod-3c618720-44b1-4fbb-8561-6aca1193652a": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034124044s - Dec 14 16:55:56.391: INFO: Pod "pod-3c618720-44b1-4fbb-8561-6aca1193652a": Phase="Pending", Reason="", readiness=false. Elapsed: 4.025732343s - Dec 14 16:55:58.392: INFO: Pod "pod-3c618720-44b1-4fbb-8561-6aca1193652a": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.027021707s - STEP: Saw pod success 12/14/22 16:55:58.392 - Dec 14 16:55:58.393: INFO: Pod "pod-3c618720-44b1-4fbb-8561-6aca1193652a" satisfied condition "Succeeded or Failed" - Dec 14 16:55:58.400: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-3c618720-44b1-4fbb-8561-6aca1193652a container test-container: - STEP: delete the pod 12/14/22 16:55:58.41 - Dec 14 16:55:58.427: INFO: Waiting for pod pod-3c618720-44b1-4fbb-8561-6aca1193652a to disappear - Dec 14 16:55:58.433: INFO: Pod pod-3c618720-44b1-4fbb-8561-6aca1193652a no longer exists - [AfterEach] [sig-storage] EmptyDir volumes + STEP: Creating a kubernetes client 02/25/23 11:27:08.542 + Feb 25 11:27:08.542: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename custom-resource-definition 02/25/23 11:27:08.545 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:08.567 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:08.571 + [It] getting/updating/patching custom resource definition status sub-resource works [Conformance] + test/e2e/apimachinery/custom_resource_definition.go:145 + Feb 25 11:27:08.577: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:55:58.433: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-3422" for this suite. 12/14/22 16:55:58.439 + Feb 25 11:27:09.354: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "custom-resource-definition-8174" for this suite. 02/25/23 11:27:09.399 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-scheduling] SchedulerPreemption [Serial] PriorityClass endpoints - verify PriorityClass endpoints can be operated with different HTTP methods [Conformance] - test/e2e/scheduling/preemption.go:733 -[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:55:58.462 -Dec 14 16:55:58.462: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename sched-preemption 12/14/22 16:55:58.464 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:58.5 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:58.504 -[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:92 -Dec 14 16:55:58.538: INFO: Waiting up to 1m0s for all nodes to be ready -Dec 14 16:56:58.581: INFO: Waiting for terminating namespaces to be deleted... -[BeforeEach] PriorityClass endpoints +[sig-node] Pods + should be submitted and removed [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:225 +[BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:56:58.586 -Dec 14 16:56:58.586: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename sched-preemption-path 12/14/22 16:56:58.59 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:56:58.624 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:56:58.627 -[BeforeEach] PriorityClass endpoints - test/e2e/scheduling/preemption.go:690 -[It] verify PriorityClass endpoints can be operated with different HTTP methods [Conformance] - test/e2e/scheduling/preemption.go:733 -Dec 14 16:56:58.653: INFO: PriorityClass.scheduling.k8s.io "p1" is invalid: value: Forbidden: may not be changed in an update. -Dec 14 16:56:58.668: INFO: PriorityClass.scheduling.k8s.io "p2" is invalid: value: Forbidden: may not be changed in an update. -[AfterEach] PriorityClass endpoints - test/e2e/framework/framework.go:187 -Dec 14 16:56:58.694: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "sched-preemption-path-7970" for this suite. 12/14/22 16:56:58.701 -[AfterEach] PriorityClass endpoints - test/e2e/scheduling/preemption.go:706 -[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] +STEP: Creating a kubernetes client 02/25/23 11:27:09.496 +Feb 25 11:27:09.496: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename pods 02/25/23 11:27:09.497 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:09.551 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:09.556 +[BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 +[It] should be submitted and removed [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:225 +STEP: creating the pod 02/25/23 11:27:09.56 +STEP: setting up watch 02/25/23 11:27:09.561 +STEP: submitting the pod to kubernetes 02/25/23 11:27:09.667 +STEP: verifying the pod is in kubernetes 02/25/23 11:27:09.725 +STEP: verifying pod creation was observed 02/25/23 11:27:09.772 +Feb 25 11:27:09.772: INFO: Waiting up to 5m0s for pod "pod-submit-remove-06589beb-8d91-4e4e-905f-e026296ef4ce" in namespace "pods-4413" to be "running" +Feb 25 11:27:09.812: INFO: Pod "pod-submit-remove-06589beb-8d91-4e4e-905f-e026296ef4ce": Phase="Pending", Reason="", readiness=false. Elapsed: 39.435698ms +Feb 25 11:27:11.819: INFO: Pod "pod-submit-remove-06589beb-8d91-4e4e-905f-e026296ef4ce": Phase="Running", Reason="", readiness=true. Elapsed: 2.046259358s +Feb 25 11:27:11.819: INFO: Pod "pod-submit-remove-06589beb-8d91-4e4e-905f-e026296ef4ce" satisfied condition "running" +STEP: deleting the pod gracefully 02/25/23 11:27:11.824 +STEP: verifying pod deletion was observed 02/25/23 11:27:11.835 +[AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 -Dec 14 16:56:58.731: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "sched-preemption-6265" for this suite. 12/14/22 16:56:58.736 -[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:80 -{"msg":"PASSED [sig-scheduling] SchedulerPreemption [Serial] PriorityClass endpoints verify PriorityClass endpoints can be operated with different HTTP methods [Conformance]","completed":293,"skipped":5645,"failed":0} +Feb 25 11:27:14.580: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-4413" for this suite. 02/25/23 11:27:14.6 +{"msg":"PASSED [sig-node] Pods should be submitted and removed [NodeConformance] [Conformance]","completed":289,"skipped":5539,"failed":0} ------------------------------ -• [SLOW TEST] [60.350 seconds] -[sig-scheduling] SchedulerPreemption [Serial] -test/e2e/scheduling/framework.go:40 - PriorityClass endpoints - test/e2e/scheduling/preemption.go:683 - verify PriorityClass endpoints can be operated with different HTTP methods [Conformance] - test/e2e/scheduling/preemption.go:733 +• [SLOW TEST] [5.121 seconds] +[sig-node] Pods +test/e2e/common/node/framework.go:23 + should be submitted and removed [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:225 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:55:58.462 - Dec 14 16:55:58.462: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename sched-preemption 12/14/22 16:55:58.464 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:55:58.5 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:55:58.504 - [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:92 - Dec 14 16:55:58.538: INFO: Waiting up to 1m0s for all nodes to be ready - Dec 14 16:56:58.581: INFO: Waiting for terminating namespaces to be deleted... - [BeforeEach] PriorityClass endpoints + [BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:56:58.586 - Dec 14 16:56:58.586: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename sched-preemption-path 12/14/22 16:56:58.59 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:56:58.624 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:56:58.627 - [BeforeEach] PriorityClass endpoints - test/e2e/scheduling/preemption.go:690 - [It] verify PriorityClass endpoints can be operated with different HTTP methods [Conformance] - test/e2e/scheduling/preemption.go:733 - Dec 14 16:56:58.653: INFO: PriorityClass.scheduling.k8s.io "p1" is invalid: value: Forbidden: may not be changed in an update. - Dec 14 16:56:58.668: INFO: PriorityClass.scheduling.k8s.io "p2" is invalid: value: Forbidden: may not be changed in an update. - [AfterEach] PriorityClass endpoints - test/e2e/framework/framework.go:187 - Dec 14 16:56:58.694: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "sched-preemption-path-7970" for this suite. 12/14/22 16:56:58.701 - [AfterEach] PriorityClass endpoints - test/e2e/scheduling/preemption.go:706 - [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] + STEP: Creating a kubernetes client 02/25/23 11:27:09.496 + Feb 25 11:27:09.496: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename pods 02/25/23 11:27:09.497 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:09.551 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:09.556 + [BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 + [It] should be submitted and removed [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:225 + STEP: creating the pod 02/25/23 11:27:09.56 + STEP: setting up watch 02/25/23 11:27:09.561 + STEP: submitting the pod to kubernetes 02/25/23 11:27:09.667 + STEP: verifying the pod is in kubernetes 02/25/23 11:27:09.725 + STEP: verifying pod creation was observed 02/25/23 11:27:09.772 + Feb 25 11:27:09.772: INFO: Waiting up to 5m0s for pod "pod-submit-remove-06589beb-8d91-4e4e-905f-e026296ef4ce" in namespace "pods-4413" to be "running" + Feb 25 11:27:09.812: INFO: Pod "pod-submit-remove-06589beb-8d91-4e4e-905f-e026296ef4ce": Phase="Pending", Reason="", readiness=false. Elapsed: 39.435698ms + Feb 25 11:27:11.819: INFO: Pod "pod-submit-remove-06589beb-8d91-4e4e-905f-e026296ef4ce": Phase="Running", Reason="", readiness=true. Elapsed: 2.046259358s + Feb 25 11:27:11.819: INFO: Pod "pod-submit-remove-06589beb-8d91-4e4e-905f-e026296ef4ce" satisfied condition "running" + STEP: deleting the pod gracefully 02/25/23 11:27:11.824 + STEP: verifying pod deletion was observed 02/25/23 11:27:11.835 + [AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 - Dec 14 16:56:58.731: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "sched-preemption-6265" for this suite. 12/14/22 16:56:58.736 - [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:80 + Feb 25 11:27:14.580: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "pods-4413" for this suite. 02/25/23 11:27:14.6 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-api-machinery] ResourceQuota - should create a ResourceQuota and capture the life of a configMap. [Conformance] - test/e2e/apimachinery/resource_quota.go:316 -[BeforeEach] [sig-api-machinery] ResourceQuota +SSSSSSSSSSSSSSS +------------------------------ +[sig-node] Sysctls [LinuxOnly] [NodeConformance] + should reject invalid sysctls [MinimumKubeletVersion:1.21] [Conformance] + test/e2e/common/node/sysctl.go:123 +[BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] + test/e2e/common/node/sysctl.go:37 +[BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:56:58.822 -Dec 14 16:56:58.823: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename resourcequota 12/14/22 16:56:58.824 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:56:58.853 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:56:58.856 -[It] should create a ResourceQuota and capture the life of a configMap. [Conformance] - test/e2e/apimachinery/resource_quota.go:316 -STEP: Counting existing ResourceQuota 12/14/22 16:57:15.867 -STEP: Creating a ResourceQuota 12/14/22 16:57:20.872 -STEP: Ensuring resource quota status is calculated 12/14/22 16:57:20.881 -STEP: Creating a ConfigMap 12/14/22 16:57:22.886 -STEP: Ensuring resource quota status captures configMap creation 12/14/22 16:57:22.9 -STEP: Deleting a ConfigMap 12/14/22 16:57:24.909 -STEP: Ensuring resource quota status released usage 12/14/22 16:57:24.92 -[AfterEach] [sig-api-machinery] ResourceQuota +STEP: Creating a kubernetes client 02/25/23 11:27:14.623 +Feb 25 11:27:14.623: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename sysctl 02/25/23 11:27:14.628 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:14.679 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:14.692 +[BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] + test/e2e/common/node/sysctl.go:67 +[It] should reject invalid sysctls [MinimumKubeletVersion:1.21] [Conformance] + test/e2e/common/node/sysctl.go:123 +STEP: Creating a pod with one valid and two invalid sysctls 02/25/23 11:27:14.698 +[AfterEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] test/e2e/framework/framework.go:187 -Dec 14 16:57:26.928: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-5983" for this suite. 12/14/22 16:57:26.933 -{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a configMap. [Conformance]","completed":294,"skipped":5665,"failed":0} +Feb 25 11:27:14.707: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "sysctl-3238" for this suite. 02/25/23 11:27:14.719 +{"msg":"PASSED [sig-node] Sysctls [LinuxOnly] [NodeConformance] should reject invalid sysctls [MinimumKubeletVersion:1.21] [Conformance]","completed":290,"skipped":5554,"failed":0} ------------------------------ -• [SLOW TEST] [28.122 seconds] -[sig-api-machinery] ResourceQuota -test/e2e/apimachinery/framework.go:23 - should create a ResourceQuota and capture the life of a configMap. [Conformance] - test/e2e/apimachinery/resource_quota.go:316 +• [0.111 seconds] +[sig-node] Sysctls [LinuxOnly] [NodeConformance] +test/e2e/common/node/framework.go:23 + should reject invalid sysctls [MinimumKubeletVersion:1.21] [Conformance] + test/e2e/common/node/sysctl.go:123 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] ResourceQuota + [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] + test/e2e/common/node/sysctl.go:37 + [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:56:58.822 - Dec 14 16:56:58.823: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename resourcequota 12/14/22 16:56:58.824 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:56:58.853 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:56:58.856 - [It] should create a ResourceQuota and capture the life of a configMap. [Conformance] - test/e2e/apimachinery/resource_quota.go:316 - STEP: Counting existing ResourceQuota 12/14/22 16:57:15.867 - STEP: Creating a ResourceQuota 12/14/22 16:57:20.872 - STEP: Ensuring resource quota status is calculated 12/14/22 16:57:20.881 - STEP: Creating a ConfigMap 12/14/22 16:57:22.886 - STEP: Ensuring resource quota status captures configMap creation 12/14/22 16:57:22.9 - STEP: Deleting a ConfigMap 12/14/22 16:57:24.909 - STEP: Ensuring resource quota status released usage 12/14/22 16:57:24.92 - [AfterEach] [sig-api-machinery] ResourceQuota + STEP: Creating a kubernetes client 02/25/23 11:27:14.623 + Feb 25 11:27:14.623: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename sysctl 02/25/23 11:27:14.628 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:14.679 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:14.692 + [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] + test/e2e/common/node/sysctl.go:67 + [It] should reject invalid sysctls [MinimumKubeletVersion:1.21] [Conformance] + test/e2e/common/node/sysctl.go:123 + STEP: Creating a pod with one valid and two invalid sysctls 02/25/23 11:27:14.698 + [AfterEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] test/e2e/framework/framework.go:187 - Dec 14 16:57:26.928: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "resourcequota-5983" for this suite. 12/14/22 16:57:26.933 + Feb 25 11:27:14.707: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "sysctl-3238" for this suite. 02/25/23 11:27:14.719 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSS +SSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] ResourceQuota - should verify ResourceQuota with terminating scopes. [Conformance] - test/e2e/apimachinery/resource_quota.go:680 -[BeforeEach] [sig-api-machinery] ResourceQuota +[sig-apps] Job + should create pods for an Indexed job with completion indexes and specified hostname [Conformance] + test/e2e/apps/job.go:194 +[BeforeEach] [sig-apps] Job test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:57:26.946 -Dec 14 16:57:26.946: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename resourcequota 12/14/22 16:57:26.952 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:57:26.981 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:57:26.985 -[It] should verify ResourceQuota with terminating scopes. [Conformance] - test/e2e/apimachinery/resource_quota.go:680 -STEP: Creating a ResourceQuota with terminating scope 12/14/22 16:57:26.991 -STEP: Ensuring ResourceQuota status is calculated 12/14/22 16:57:26.997 -STEP: Creating a ResourceQuota with not terminating scope 12/14/22 16:57:29.002 -STEP: Ensuring ResourceQuota status is calculated 12/14/22 16:57:29.009 -STEP: Creating a long running pod 12/14/22 16:57:31.022 -STEP: Ensuring resource quota with not terminating scope captures the pod usage 12/14/22 16:57:31.044 -STEP: Ensuring resource quota with terminating scope ignored the pod usage 12/14/22 16:57:33.052 -STEP: Deleting the pod 12/14/22 16:57:35.062 -STEP: Ensuring resource quota status released the pod usage 12/14/22 16:57:35.083 -STEP: Creating a terminating pod 12/14/22 16:57:37.105 -STEP: Ensuring resource quota with terminating scope captures the pod usage 12/14/22 16:57:37.149 -STEP: Ensuring resource quota with not terminating scope ignored the pod usage 12/14/22 16:57:39.156 -STEP: Deleting the pod 12/14/22 16:57:41.164 -STEP: Ensuring resource quota status released the pod usage 12/14/22 16:57:41.181 -[AfterEach] [sig-api-machinery] ResourceQuota +STEP: Creating a kubernetes client 02/25/23 11:27:14.747 +Feb 25 11:27:14.751: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename job 02/25/23 11:27:14.763 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:14.793 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:14.799 +[It] should create pods for an Indexed job with completion indexes and specified hostname [Conformance] + test/e2e/apps/job.go:194 +STEP: Creating Indexed job 02/25/23 11:27:14.803 +STEP: Ensuring job reaches completions 02/25/23 11:27:14.826 +STEP: Ensuring pods with index for job exist 02/25/23 11:27:24.831 +[AfterEach] [sig-apps] Job test/e2e/framework/framework.go:187 -Dec 14 16:57:43.186: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-5330" for this suite. 12/14/22 16:57:43.191 -{"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with terminating scopes. [Conformance]","completed":295,"skipped":5672,"failed":0} +Feb 25 11:27:24.837: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "job-6071" for this suite. 02/25/23 11:27:24.845 +{"msg":"PASSED [sig-apps] Job should create pods for an Indexed job with completion indexes and specified hostname [Conformance]","completed":291,"skipped":5571,"failed":0} ------------------------------ -• [SLOW TEST] [16.253 seconds] -[sig-api-machinery] ResourceQuota -test/e2e/apimachinery/framework.go:23 - should verify ResourceQuota with terminating scopes. [Conformance] - test/e2e/apimachinery/resource_quota.go:680 +• [SLOW TEST] [10.189 seconds] +[sig-apps] Job +test/e2e/apps/framework.go:23 + should create pods for an Indexed job with completion indexes and specified hostname [Conformance] + test/e2e/apps/job.go:194 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] ResourceQuota + [BeforeEach] [sig-apps] Job test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:57:26.946 - Dec 14 16:57:26.946: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename resourcequota 12/14/22 16:57:26.952 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:57:26.981 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:57:26.985 - [It] should verify ResourceQuota with terminating scopes. [Conformance] - test/e2e/apimachinery/resource_quota.go:680 - STEP: Creating a ResourceQuota with terminating scope 12/14/22 16:57:26.991 - STEP: Ensuring ResourceQuota status is calculated 12/14/22 16:57:26.997 - STEP: Creating a ResourceQuota with not terminating scope 12/14/22 16:57:29.002 - STEP: Ensuring ResourceQuota status is calculated 12/14/22 16:57:29.009 - STEP: Creating a long running pod 12/14/22 16:57:31.022 - STEP: Ensuring resource quota with not terminating scope captures the pod usage 12/14/22 16:57:31.044 - STEP: Ensuring resource quota with terminating scope ignored the pod usage 12/14/22 16:57:33.052 - STEP: Deleting the pod 12/14/22 16:57:35.062 - STEP: Ensuring resource quota status released the pod usage 12/14/22 16:57:35.083 - STEP: Creating a terminating pod 12/14/22 16:57:37.105 - STEP: Ensuring resource quota with terminating scope captures the pod usage 12/14/22 16:57:37.149 - STEP: Ensuring resource quota with not terminating scope ignored the pod usage 12/14/22 16:57:39.156 - STEP: Deleting the pod 12/14/22 16:57:41.164 - STEP: Ensuring resource quota status released the pod usage 12/14/22 16:57:41.181 - [AfterEach] [sig-api-machinery] ResourceQuota + STEP: Creating a kubernetes client 02/25/23 11:27:14.747 + Feb 25 11:27:14.751: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename job 02/25/23 11:27:14.763 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:14.793 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:14.799 + [It] should create pods for an Indexed job with completion indexes and specified hostname [Conformance] + test/e2e/apps/job.go:194 + STEP: Creating Indexed job 02/25/23 11:27:14.803 + STEP: Ensuring job reaches completions 02/25/23 11:27:14.826 + STEP: Ensuring pods with index for job exist 02/25/23 11:27:24.831 + [AfterEach] [sig-apps] Job test/e2e/framework/framework.go:187 - Dec 14 16:57:43.186: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "resourcequota-5330" for this suite. 12/14/22 16:57:43.191 + Feb 25 11:27:24.837: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "job-6071" for this suite. 02/25/23 11:27:24.845 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] Garbage collector - should delete RS created by deployment when not orphaning [Conformance] - test/e2e/apimachinery/garbage_collector.go:491 -[BeforeEach] [sig-api-machinery] Garbage collector +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + updates the published spec when one version gets renamed [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:390 +[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:57:43.2 -Dec 14 16:57:43.200: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename gc 12/14/22 16:57:43.202 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:57:43.239 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:57:43.243 -[It] should delete RS created by deployment when not orphaning [Conformance] - test/e2e/apimachinery/garbage_collector.go:491 -STEP: create the deployment 12/14/22 16:57:43.247 -STEP: Wait for the Deployment to create new ReplicaSet 12/14/22 16:57:43.257 -STEP: delete the deployment 12/14/22 16:57:43.265 -STEP: wait for all rs to be garbage collected 12/14/22 16:57:43.277 -STEP: expected 0 rs, got 1 rs 12/14/22 16:57:43.287 -STEP: expected 0 pods, got 1 pods 12/14/22 16:57:43.291 -STEP: Gathering metrics 12/14/22 16:57:43.808 -Dec 14 16:57:43.997: INFO: Waiting up to 5m0s for pod "kube-controller-manager-kiedai4iakae-2" in namespace "kube-system" to be "running and ready" -Dec 14 16:57:44.006: INFO: Pod "kube-controller-manager-kiedai4iakae-2": Phase="Running", Reason="", readiness=true. Elapsed: 8.985564ms -Dec 14 16:57:44.006: INFO: The phase of Pod kube-controller-manager-kiedai4iakae-2 is Running (Ready = true) -Dec 14 16:57:44.006: INFO: Pod "kube-controller-manager-kiedai4iakae-2" satisfied condition "running and ready" -Dec 14 16:57:44.112: INFO: For apiserver_request_total: -For apiserver_request_latency_seconds: -For apiserver_init_events_total: -For garbage_collector_attempt_to_delete_queue_latency: -For garbage_collector_attempt_to_delete_work_duration: -For garbage_collector_attempt_to_orphan_queue_latency: -For garbage_collector_attempt_to_orphan_work_duration: -For garbage_collector_dirty_processing_latency_microseconds: -For garbage_collector_event_processing_latency_microseconds: -For garbage_collector_graph_changes_queue_latency: -For garbage_collector_graph_changes_work_duration: -For garbage_collector_orphan_processing_latency_microseconds: -For namespace_queue_latency: -For namespace_queue_latency_sum: -For namespace_queue_latency_count: -For namespace_retries: -For namespace_work_duration: -For namespace_work_duration_sum: -For namespace_work_duration_count: -For function_duration_seconds: -For errors_total: -For evicted_pods_total: - -[AfterEach] [sig-api-machinery] Garbage collector +STEP: Creating a kubernetes client 02/25/23 11:27:24.946 +Feb 25 11:27:24.946: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 11:27:24.952 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:24.975 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:24.978 +[It] updates the published spec when one version gets renamed [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:390 +STEP: set up a multi version CRD 02/25/23 11:27:24.982 +Feb 25 11:27:24.983: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: rename a version 02/25/23 11:27:39.711 +STEP: check the new version name is served 02/25/23 11:27:39.733 +STEP: check the old version name is removed 02/25/23 11:27:46.007 +STEP: check the other version is not changed 02/25/23 11:27:48.205 +[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:57:44.113: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "gc-1281" for this suite. 12/14/22 16:57:44.12 -{"msg":"PASSED [sig-api-machinery] Garbage collector should delete RS created by deployment when not orphaning [Conformance]","completed":296,"skipped":5688,"failed":0} +Feb 25 11:27:58.178: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "crd-publish-openapi-2713" for this suite. 02/25/23 11:27:58.196 +{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] updates the published spec when one version gets renamed [Conformance]","completed":292,"skipped":5609,"failed":0} ------------------------------ -• [0.935 seconds] -[sig-api-machinery] Garbage collector +• [SLOW TEST] [33.260 seconds] +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/apimachinery/framework.go:23 - should delete RS created by deployment when not orphaning [Conformance] - test/e2e/apimachinery/garbage_collector.go:491 + updates the published spec when one version gets renamed [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:390 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Garbage collector + [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:57:43.2 - Dec 14 16:57:43.200: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename gc 12/14/22 16:57:43.202 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:57:43.239 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:57:43.243 - [It] should delete RS created by deployment when not orphaning [Conformance] - test/e2e/apimachinery/garbage_collector.go:491 - STEP: create the deployment 12/14/22 16:57:43.247 - STEP: Wait for the Deployment to create new ReplicaSet 12/14/22 16:57:43.257 - STEP: delete the deployment 12/14/22 16:57:43.265 - STEP: wait for all rs to be garbage collected 12/14/22 16:57:43.277 - STEP: expected 0 rs, got 1 rs 12/14/22 16:57:43.287 - STEP: expected 0 pods, got 1 pods 12/14/22 16:57:43.291 - STEP: Gathering metrics 12/14/22 16:57:43.808 - Dec 14 16:57:43.997: INFO: Waiting up to 5m0s for pod "kube-controller-manager-kiedai4iakae-2" in namespace "kube-system" to be "running and ready" - Dec 14 16:57:44.006: INFO: Pod "kube-controller-manager-kiedai4iakae-2": Phase="Running", Reason="", readiness=true. Elapsed: 8.985564ms - Dec 14 16:57:44.006: INFO: The phase of Pod kube-controller-manager-kiedai4iakae-2 is Running (Ready = true) - Dec 14 16:57:44.006: INFO: Pod "kube-controller-manager-kiedai4iakae-2" satisfied condition "running and ready" - Dec 14 16:57:44.112: INFO: For apiserver_request_total: - For apiserver_request_latency_seconds: - For apiserver_init_events_total: - For garbage_collector_attempt_to_delete_queue_latency: - For garbage_collector_attempt_to_delete_work_duration: - For garbage_collector_attempt_to_orphan_queue_latency: - For garbage_collector_attempt_to_orphan_work_duration: - For garbage_collector_dirty_processing_latency_microseconds: - For garbage_collector_event_processing_latency_microseconds: - For garbage_collector_graph_changes_queue_latency: - For garbage_collector_graph_changes_work_duration: - For garbage_collector_orphan_processing_latency_microseconds: - For namespace_queue_latency: - For namespace_queue_latency_sum: - For namespace_queue_latency_count: - For namespace_retries: - For namespace_work_duration: - For namespace_work_duration_sum: - For namespace_work_duration_count: - For function_duration_seconds: - For errors_total: - For evicted_pods_total: - - [AfterEach] [sig-api-machinery] Garbage collector + STEP: Creating a kubernetes client 02/25/23 11:27:24.946 + Feb 25 11:27:24.946: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 11:27:24.952 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:24.975 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:24.978 + [It] updates the published spec when one version gets renamed [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:390 + STEP: set up a multi version CRD 02/25/23 11:27:24.982 + Feb 25 11:27:24.983: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: rename a version 02/25/23 11:27:39.711 + STEP: check the new version name is served 02/25/23 11:27:39.733 + STEP: check the old version name is removed 02/25/23 11:27:46.007 + STEP: check the other version is not changed 02/25/23 11:27:48.205 + [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:57:44.113: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "gc-1281" for this suite. 12/14/22 16:57:44.12 + Feb 25 11:27:58.178: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "crd-publish-openapi-2713" for this suite. 02/25/23 11:27:58.196 << End Captured GinkgoWriter Output ------------------------------ -[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition - creating/deleting custom resource definition objects works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:58 -[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +SSSSSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-api-machinery] Namespaces [Serial] + should patch a Namespace [Conformance] + test/e2e/apimachinery/namespace.go:267 +[BeforeEach] [sig-api-machinery] Namespaces [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:57:44.137 -Dec 14 16:57:44.137: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename custom-resource-definition 12/14/22 16:57:44.14 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:57:44.167 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:57:44.17 -[It] creating/deleting custom resource definition objects works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:58 -Dec 14 16:57:44.173: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 11:27:58.209 +Feb 25 11:27:58.209: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename namespaces 02/25/23 11:27:58.214 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:58.242 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:58.246 +[It] should patch a Namespace [Conformance] + test/e2e/apimachinery/namespace.go:267 +STEP: creating a Namespace 02/25/23 11:27:58.25 +STEP: patching the Namespace 02/25/23 11:27:58.274 +STEP: get the Namespace and ensuring it has the label 02/25/23 11:27:58.283 +[AfterEach] [sig-api-machinery] Namespaces [Serial] test/e2e/framework/framework.go:187 -Dec 14 16:57:45.210: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "custom-resource-definition-7500" for this suite. 12/14/22 16:57:45.218 -{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] Simple CustomResourceDefinition creating/deleting custom resource definition objects works [Conformance]","completed":297,"skipped":5688,"failed":0} +Feb 25 11:27:58.289: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "namespaces-8619" for this suite. 02/25/23 11:27:58.295 +STEP: Destroying namespace "nspatchtest-40f556a9-ed0e-43b3-8a71-45f81c05a841-7491" for this suite. 02/25/23 11:27:58.305 +{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should patch a Namespace [Conformance]","completed":293,"skipped":5631,"failed":0} ------------------------------ -• [1.089 seconds] -[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +• [0.106 seconds] +[sig-api-machinery] Namespaces [Serial] test/e2e/apimachinery/framework.go:23 - Simple CustomResourceDefinition - test/e2e/apimachinery/custom_resource_definition.go:50 - creating/deleting custom resource definition objects works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:58 + should patch a Namespace [Conformance] + test/e2e/apimachinery/namespace.go:267 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] + [BeforeEach] [sig-api-machinery] Namespaces [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:57:44.137 - Dec 14 16:57:44.137: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename custom-resource-definition 12/14/22 16:57:44.14 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:57:44.167 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:57:44.17 - [It] creating/deleting custom resource definition objects works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:58 - Dec 14 16:57:44.173: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 11:27:58.209 + Feb 25 11:27:58.209: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename namespaces 02/25/23 11:27:58.214 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:58.242 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:58.246 + [It] should patch a Namespace [Conformance] + test/e2e/apimachinery/namespace.go:267 + STEP: creating a Namespace 02/25/23 11:27:58.25 + STEP: patching the Namespace 02/25/23 11:27:58.274 + STEP: get the Namespace and ensuring it has the label 02/25/23 11:27:58.283 + [AfterEach] [sig-api-machinery] Namespaces [Serial] test/e2e/framework/framework.go:187 - Dec 14 16:57:45.210: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "custom-resource-definition-7500" for this suite. 12/14/22 16:57:45.218 + Feb 25 11:27:58.289: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "namespaces-8619" for this suite. 02/25/23 11:27:58.295 + STEP: Destroying namespace "nspatchtest-40f556a9-ed0e-43b3-8a71-45f81c05a841-7491" for this suite. 02/25/23 11:27:58.305 << End Captured GinkgoWriter Output ------------------------------ -S +SSSSSSS ------------------------------ -[sig-apps] ControllerRevision [Serial] - should manage the lifecycle of a ControllerRevision [Conformance] - test/e2e/apps/controller_revision.go:124 -[BeforeEach] [sig-apps] ControllerRevision [Serial] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + patching/updating a mutating webhook should work [Conformance] + test/e2e/apimachinery/webhook.go:507 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:57:45.228 -Dec 14 16:57:45.228: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename controllerrevisions 12/14/22 16:57:45.231 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:57:45.26 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:57:45.263 -[BeforeEach] [sig-apps] ControllerRevision [Serial] - test/e2e/apps/controller_revision.go:93 -[It] should manage the lifecycle of a ControllerRevision [Conformance] - test/e2e/apps/controller_revision.go:124 -STEP: Creating DaemonSet "e2e-w2zpb-daemon-set" 12/14/22 16:57:45.293 -STEP: Check that daemon pods launch on every node of the cluster. 12/14/22 16:57:45.301 -Dec 14 16:57:45.319: INFO: Number of nodes with available pods controlled by daemonset e2e-w2zpb-daemon-set: 0 -Dec 14 16:57:45.319: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 16:57:46.335: INFO: Number of nodes with available pods controlled by daemonset e2e-w2zpb-daemon-set: 0 -Dec 14 16:57:46.335: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 16:57:47.331: INFO: Number of nodes with available pods controlled by daemonset e2e-w2zpb-daemon-set: 3 -Dec 14 16:57:47.332: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset e2e-w2zpb-daemon-set -STEP: Confirm DaemonSet "e2e-w2zpb-daemon-set" successfully created with "daemonset-name=e2e-w2zpb-daemon-set" label 12/14/22 16:57:47.337 -STEP: Listing all ControllerRevisions with label "daemonset-name=e2e-w2zpb-daemon-set" 12/14/22 16:57:47.347 -Dec 14 16:57:47.352: INFO: Located ControllerRevision: "e2e-w2zpb-daemon-set-75fbdc6c6d" -STEP: Patching ControllerRevision "e2e-w2zpb-daemon-set-75fbdc6c6d" 12/14/22 16:57:47.356 -Dec 14 16:57:47.366: INFO: e2e-w2zpb-daemon-set-75fbdc6c6d has been patched -STEP: Create a new ControllerRevision 12/14/22 16:57:47.366 -Dec 14 16:57:47.373: INFO: Created ControllerRevision: e2e-w2zpb-daemon-set-55f9b9dc86 -STEP: Confirm that there are two ControllerRevisions 12/14/22 16:57:47.373 -Dec 14 16:57:47.373: INFO: Requesting list of ControllerRevisions to confirm quantity -Dec 14 16:57:47.377: INFO: Found 2 ControllerRevisions -STEP: Deleting ControllerRevision "e2e-w2zpb-daemon-set-75fbdc6c6d" 12/14/22 16:57:47.377 -STEP: Confirm that there is only one ControllerRevision 12/14/22 16:57:47.384 -Dec 14 16:57:47.384: INFO: Requesting list of ControllerRevisions to confirm quantity -Dec 14 16:57:47.387: INFO: Found 1 ControllerRevisions -STEP: Updating ControllerRevision "e2e-w2zpb-daemon-set-55f9b9dc86" 12/14/22 16:57:47.39 -Dec 14 16:57:47.399: INFO: e2e-w2zpb-daemon-set-55f9b9dc86 has been updated -STEP: Generate another ControllerRevision by patching the Daemonset 12/14/22 16:57:47.399 -W1214 16:57:47.406472 14 warnings.go:70] unknown field "updateStrategy" -STEP: Confirm that there are two ControllerRevisions 12/14/22 16:57:47.406 -Dec 14 16:57:47.406: INFO: Requesting list of ControllerRevisions to confirm quantity -Dec 14 16:57:48.409: INFO: Requesting list of ControllerRevisions to confirm quantity -Dec 14 16:57:48.415: INFO: Found 2 ControllerRevisions -STEP: Removing a ControllerRevision via 'DeleteCollection' with labelSelector: "e2e-w2zpb-daemon-set-55f9b9dc86=updated" 12/14/22 16:57:48.415 -STEP: Confirm that there is only one ControllerRevision 12/14/22 16:57:48.427 -Dec 14 16:57:48.427: INFO: Requesting list of ControllerRevisions to confirm quantity -Dec 14 16:57:48.433: INFO: Found 1 ControllerRevisions -Dec 14 16:57:48.438: INFO: ControllerRevision "e2e-w2zpb-daemon-set-78b59f6979" has revision 3 -[AfterEach] [sig-apps] ControllerRevision [Serial] - test/e2e/apps/controller_revision.go:58 -STEP: Deleting DaemonSet "e2e-w2zpb-daemon-set" 12/14/22 16:57:48.443 -STEP: deleting DaemonSet.extensions e2e-w2zpb-daemon-set in namespace controllerrevisions-6566, will wait for the garbage collector to delete the pods 12/14/22 16:57:48.443 -Dec 14 16:57:48.514: INFO: Deleting DaemonSet.extensions e2e-w2zpb-daemon-set took: 9.907153ms -Dec 14 16:57:48.616: INFO: Terminating DaemonSet.extensions e2e-w2zpb-daemon-set pods took: 101.194691ms -Dec 14 16:57:50.221: INFO: Number of nodes with available pods controlled by daemonset e2e-w2zpb-daemon-set: 0 -Dec 14 16:57:50.221: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset e2e-w2zpb-daemon-set -Dec 14 16:57:50.227: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"28149"},"items":null} - -Dec 14 16:57:50.232: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"28149"},"items":null} - -[AfterEach] [sig-apps] ControllerRevision [Serial] +STEP: Creating a kubernetes client 02/25/23 11:27:58.316 +Feb 25 11:27:58.317: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 11:27:58.32 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:58.341 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:58.345 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 +STEP: Setting up server cert 02/25/23 11:27:58.372 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 11:27:59.03 +STEP: Deploying the webhook pod 02/25/23 11:27:59.046 +STEP: Wait for the deployment to be ready 02/25/23 11:27:59.067 +Feb 25 11:27:59.083: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set +STEP: Deploying the webhook service 02/25/23 11:28:01.181 +STEP: Verifying the service has paired with the endpoint 02/25/23 11:28:01.246 +Feb 25 11:28:02.247: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] patching/updating a mutating webhook should work [Conformance] + test/e2e/apimachinery/webhook.go:507 +STEP: Creating a mutating webhook configuration 02/25/23 11:28:02.254 +STEP: Updating a mutating webhook configuration's rules to not include the create operation 02/25/23 11:28:02.293 +STEP: Creating a configMap that should not be mutated 02/25/23 11:28:02.304 +STEP: Patching a mutating webhook configuration's rules to include the create operation 02/25/23 11:28:02.322 +STEP: Creating a configMap that should be mutated 02/25/23 11:28:02.332 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:57:50.250: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "controllerrevisions-6566" for this suite. 12/14/22 16:57:50.256 -{"msg":"PASSED [sig-apps] ControllerRevision [Serial] should manage the lifecycle of a ControllerRevision [Conformance]","completed":298,"skipped":5689,"failed":0} +Feb 25 11:28:02.372: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-2957" for this suite. 02/25/23 11:28:02.382 +STEP: Destroying namespace "webhook-2957-markers" for this suite. 02/25/23 11:28:02.395 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a mutating webhook should work [Conformance]","completed":294,"skipped":5638,"failed":0} ------------------------------ -• [SLOW TEST] [5.036 seconds] -[sig-apps] ControllerRevision [Serial] -test/e2e/apps/framework.go:23 - should manage the lifecycle of a ControllerRevision [Conformance] - test/e2e/apps/controller_revision.go:124 +• [4.184 seconds] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + patching/updating a mutating webhook should work [Conformance] + test/e2e/apimachinery/webhook.go:507 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] ControllerRevision [Serial] + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:57:45.228 - Dec 14 16:57:45.228: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename controllerrevisions 12/14/22 16:57:45.231 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:57:45.26 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:57:45.263 - [BeforeEach] [sig-apps] ControllerRevision [Serial] - test/e2e/apps/controller_revision.go:93 - [It] should manage the lifecycle of a ControllerRevision [Conformance] - test/e2e/apps/controller_revision.go:124 - STEP: Creating DaemonSet "e2e-w2zpb-daemon-set" 12/14/22 16:57:45.293 - STEP: Check that daemon pods launch on every node of the cluster. 12/14/22 16:57:45.301 - Dec 14 16:57:45.319: INFO: Number of nodes with available pods controlled by daemonset e2e-w2zpb-daemon-set: 0 - Dec 14 16:57:45.319: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 16:57:46.335: INFO: Number of nodes with available pods controlled by daemonset e2e-w2zpb-daemon-set: 0 - Dec 14 16:57:46.335: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 16:57:47.331: INFO: Number of nodes with available pods controlled by daemonset e2e-w2zpb-daemon-set: 3 - Dec 14 16:57:47.332: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset e2e-w2zpb-daemon-set - STEP: Confirm DaemonSet "e2e-w2zpb-daemon-set" successfully created with "daemonset-name=e2e-w2zpb-daemon-set" label 12/14/22 16:57:47.337 - STEP: Listing all ControllerRevisions with label "daemonset-name=e2e-w2zpb-daemon-set" 12/14/22 16:57:47.347 - Dec 14 16:57:47.352: INFO: Located ControllerRevision: "e2e-w2zpb-daemon-set-75fbdc6c6d" - STEP: Patching ControllerRevision "e2e-w2zpb-daemon-set-75fbdc6c6d" 12/14/22 16:57:47.356 - Dec 14 16:57:47.366: INFO: e2e-w2zpb-daemon-set-75fbdc6c6d has been patched - STEP: Create a new ControllerRevision 12/14/22 16:57:47.366 - Dec 14 16:57:47.373: INFO: Created ControllerRevision: e2e-w2zpb-daemon-set-55f9b9dc86 - STEP: Confirm that there are two ControllerRevisions 12/14/22 16:57:47.373 - Dec 14 16:57:47.373: INFO: Requesting list of ControllerRevisions to confirm quantity - Dec 14 16:57:47.377: INFO: Found 2 ControllerRevisions - STEP: Deleting ControllerRevision "e2e-w2zpb-daemon-set-75fbdc6c6d" 12/14/22 16:57:47.377 - STEP: Confirm that there is only one ControllerRevision 12/14/22 16:57:47.384 - Dec 14 16:57:47.384: INFO: Requesting list of ControllerRevisions to confirm quantity - Dec 14 16:57:47.387: INFO: Found 1 ControllerRevisions - STEP: Updating ControllerRevision "e2e-w2zpb-daemon-set-55f9b9dc86" 12/14/22 16:57:47.39 - Dec 14 16:57:47.399: INFO: e2e-w2zpb-daemon-set-55f9b9dc86 has been updated - STEP: Generate another ControllerRevision by patching the Daemonset 12/14/22 16:57:47.399 - W1214 16:57:47.406472 14 warnings.go:70] unknown field "updateStrategy" - STEP: Confirm that there are two ControllerRevisions 12/14/22 16:57:47.406 - Dec 14 16:57:47.406: INFO: Requesting list of ControllerRevisions to confirm quantity - Dec 14 16:57:48.409: INFO: Requesting list of ControllerRevisions to confirm quantity - Dec 14 16:57:48.415: INFO: Found 2 ControllerRevisions - STEP: Removing a ControllerRevision via 'DeleteCollection' with labelSelector: "e2e-w2zpb-daemon-set-55f9b9dc86=updated" 12/14/22 16:57:48.415 - STEP: Confirm that there is only one ControllerRevision 12/14/22 16:57:48.427 - Dec 14 16:57:48.427: INFO: Requesting list of ControllerRevisions to confirm quantity - Dec 14 16:57:48.433: INFO: Found 1 ControllerRevisions - Dec 14 16:57:48.438: INFO: ControllerRevision "e2e-w2zpb-daemon-set-78b59f6979" has revision 3 - [AfterEach] [sig-apps] ControllerRevision [Serial] - test/e2e/apps/controller_revision.go:58 - STEP: Deleting DaemonSet "e2e-w2zpb-daemon-set" 12/14/22 16:57:48.443 - STEP: deleting DaemonSet.extensions e2e-w2zpb-daemon-set in namespace controllerrevisions-6566, will wait for the garbage collector to delete the pods 12/14/22 16:57:48.443 - Dec 14 16:57:48.514: INFO: Deleting DaemonSet.extensions e2e-w2zpb-daemon-set took: 9.907153ms - Dec 14 16:57:48.616: INFO: Terminating DaemonSet.extensions e2e-w2zpb-daemon-set pods took: 101.194691ms - Dec 14 16:57:50.221: INFO: Number of nodes with available pods controlled by daemonset e2e-w2zpb-daemon-set: 0 - Dec 14 16:57:50.221: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset e2e-w2zpb-daemon-set - Dec 14 16:57:50.227: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"28149"},"items":null} - - Dec 14 16:57:50.232: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"28149"},"items":null} - - [AfterEach] [sig-apps] ControllerRevision [Serial] + STEP: Creating a kubernetes client 02/25/23 11:27:58.316 + Feb 25 11:27:58.317: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 11:27:58.32 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:27:58.341 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:27:58.345 + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 + STEP: Setting up server cert 02/25/23 11:27:58.372 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 11:27:59.03 + STEP: Deploying the webhook pod 02/25/23 11:27:59.046 + STEP: Wait for the deployment to be ready 02/25/23 11:27:59.067 + Feb 25 11:27:59.083: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set + STEP: Deploying the webhook service 02/25/23 11:28:01.181 + STEP: Verifying the service has paired with the endpoint 02/25/23 11:28:01.246 + Feb 25 11:28:02.247: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] patching/updating a mutating webhook should work [Conformance] + test/e2e/apimachinery/webhook.go:507 + STEP: Creating a mutating webhook configuration 02/25/23 11:28:02.254 + STEP: Updating a mutating webhook configuration's rules to not include the create operation 02/25/23 11:28:02.293 + STEP: Creating a configMap that should not be mutated 02/25/23 11:28:02.304 + STEP: Patching a mutating webhook configuration's rules to include the create operation 02/25/23 11:28:02.322 + STEP: Creating a configMap that should be mutated 02/25/23 11:28:02.332 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:57:50.250: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "controllerrevisions-6566" for this suite. 12/14/22 16:57:50.256 + Feb 25 11:28:02.372: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-2957" for this suite. 02/25/23 11:28:02.382 + STEP: Destroying namespace "webhook-2957-markers" for this suite. 02/25/23 11:28:02.395 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSS ------------------------------- -[sig-storage] Projected secret - should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:87 -[BeforeEach] [sig-storage] Projected secret +[sig-storage] Projected configMap + should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:46 +[BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:57:50.269 -Dec 14 16:57:50.270: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 16:57:50.272 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:57:50.293 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:57:50.296 -[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:87 -STEP: Creating projection with secret that has name projected-secret-test-map-fe98012c-bd88-487e-89bb-f214ba27b8e0 12/14/22 16:57:50.3 -STEP: Creating a pod to test consume secrets 12/14/22 16:57:50.307 -Dec 14 16:57:50.317: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc" in namespace "projected-3624" to be "Succeeded or Failed" -Dec 14 16:57:50.322: INFO: Pod "pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc": Phase="Pending", Reason="", readiness=false. Elapsed: 4.958896ms -Dec 14 16:57:52.336: INFO: Pod "pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019015665s -Dec 14 16:57:54.330: INFO: Pod "pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013806208s -STEP: Saw pod success 12/14/22 16:57:54.331 -Dec 14 16:57:54.331: INFO: Pod "pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc" satisfied condition "Succeeded or Failed" -Dec 14 16:57:54.336: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc container projected-secret-volume-test: -STEP: delete the pod 12/14/22 16:57:54.355 -Dec 14 16:57:54.368: INFO: Waiting for pod pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc to disappear -Dec 14 16:57:54.372: INFO: Pod pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc no longer exists -[AfterEach] [sig-storage] Projected secret +STEP: Creating a kubernetes client 02/25/23 11:28:02.504 +Feb 25 11:28:02.505: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 11:28:02.508 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:02.553 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:02.559 +[It] should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:46 +STEP: Creating configMap with name projected-configmap-test-volume-115c58f9-1668-4da3-a49a-020b25f40d91 02/25/23 11:28:02.563 +STEP: Creating a pod to test consume configMaps 02/25/23 11:28:02.571 +Feb 25 11:28:02.599: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623" in namespace "projected-5594" to be "Succeeded or Failed" +Feb 25 11:28:02.610: INFO: Pod "pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623": Phase="Pending", Reason="", readiness=false. Elapsed: 10.793245ms +Feb 25 11:28:04.624: INFO: Pod "pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623": Phase="Running", Reason="", readiness=true. Elapsed: 2.025225642s +Feb 25 11:28:06.617: INFO: Pod "pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623": Phase="Running", Reason="", readiness=false. Elapsed: 4.018302024s +Feb 25 11:28:08.617: INFO: Pod "pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623": Phase="Running", Reason="", readiness=false. Elapsed: 6.018143474s +Feb 25 11:28:10.619: INFO: Pod "pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.019889829s +STEP: Saw pod success 02/25/23 11:28:10.619 +Feb 25 11:28:10.619: INFO: Pod "pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623" satisfied condition "Succeeded or Failed" +Feb 25 11:28:10.623: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623 container agnhost-container: +STEP: delete the pod 02/25/23 11:28:10.646 +Feb 25 11:28:10.665: INFO: Waiting for pod pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623 to disappear +Feb 25 11:28:10.671: INFO: Pod pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623 no longer exists +[AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 -Dec 14 16:57:54.372: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-3624" for this suite. 12/14/22 16:57:54.376 -{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","completed":299,"skipped":5700,"failed":0} +Feb 25 11:28:10.671: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-5594" for this suite. 02/25/23 11:28:10.677 +{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume [NodeConformance] [Conformance]","completed":295,"skipped":5638,"failed":0} ------------------------------ -• [4.118 seconds] -[sig-storage] Projected secret +• [SLOW TEST] [8.183 seconds] +[sig-storage] Projected configMap test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:87 + should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:46 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected secret + [BeforeEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:57:50.269 - Dec 14 16:57:50.270: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 16:57:50.272 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:57:50.293 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:57:50.296 - [It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_secret.go:87 - STEP: Creating projection with secret that has name projected-secret-test-map-fe98012c-bd88-487e-89bb-f214ba27b8e0 12/14/22 16:57:50.3 - STEP: Creating a pod to test consume secrets 12/14/22 16:57:50.307 - Dec 14 16:57:50.317: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc" in namespace "projected-3624" to be "Succeeded or Failed" - Dec 14 16:57:50.322: INFO: Pod "pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc": Phase="Pending", Reason="", readiness=false. Elapsed: 4.958896ms - Dec 14 16:57:52.336: INFO: Pod "pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019015665s - Dec 14 16:57:54.330: INFO: Pod "pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013806208s - STEP: Saw pod success 12/14/22 16:57:54.331 - Dec 14 16:57:54.331: INFO: Pod "pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc" satisfied condition "Succeeded or Failed" - Dec 14 16:57:54.336: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc container projected-secret-volume-test: - STEP: delete the pod 12/14/22 16:57:54.355 - Dec 14 16:57:54.368: INFO: Waiting for pod pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc to disappear - Dec 14 16:57:54.372: INFO: Pod pod-projected-secrets-b6199b72-15a7-46d6-b320-6d6beea647bc no longer exists - [AfterEach] [sig-storage] Projected secret + STEP: Creating a kubernetes client 02/25/23 11:28:02.504 + Feb 25 11:28:02.505: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 11:28:02.508 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:02.553 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:02.559 + [It] should be consumable from pods in volume [NodeConformance] [Conformance] + test/e2e/common/storage/projected_configmap.go:46 + STEP: Creating configMap with name projected-configmap-test-volume-115c58f9-1668-4da3-a49a-020b25f40d91 02/25/23 11:28:02.563 + STEP: Creating a pod to test consume configMaps 02/25/23 11:28:02.571 + Feb 25 11:28:02.599: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623" in namespace "projected-5594" to be "Succeeded or Failed" + Feb 25 11:28:02.610: INFO: Pod "pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623": Phase="Pending", Reason="", readiness=false. Elapsed: 10.793245ms + Feb 25 11:28:04.624: INFO: Pod "pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623": Phase="Running", Reason="", readiness=true. Elapsed: 2.025225642s + Feb 25 11:28:06.617: INFO: Pod "pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623": Phase="Running", Reason="", readiness=false. Elapsed: 4.018302024s + Feb 25 11:28:08.617: INFO: Pod "pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623": Phase="Running", Reason="", readiness=false. Elapsed: 6.018143474s + Feb 25 11:28:10.619: INFO: Pod "pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.019889829s + STEP: Saw pod success 02/25/23 11:28:10.619 + Feb 25 11:28:10.619: INFO: Pod "pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623" satisfied condition "Succeeded or Failed" + Feb 25 11:28:10.623: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623 container agnhost-container: + STEP: delete the pod 02/25/23 11:28:10.646 + Feb 25 11:28:10.665: INFO: Waiting for pod pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623 to disappear + Feb 25 11:28:10.671: INFO: Pod pod-projected-configmaps-5eaa0b77-c0d5-48b5-afe6-90e8e076f623 no longer exists + [AfterEach] [sig-storage] Projected configMap test/e2e/framework/framework.go:187 - Dec 14 16:57:54.372: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-3624" for this suite. 12/14/22 16:57:54.376 + Feb 25 11:28:10.671: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-5594" for this suite. 02/25/23 11:28:10.677 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSS +SSSSSSSS ------------------------------ -[sig-node] Security Context When creating a container with runAsUser - should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:346 -[BeforeEach] [sig-node] Security Context +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + should mutate custom resource [Conformance] + test/e2e/apimachinery/webhook.go:290 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:57:54.388 -Dec 14 16:57:54.388: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename security-context-test 12/14/22 16:57:54.392 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:57:54.423 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:57:54.427 -[BeforeEach] [sig-node] Security Context - test/e2e/common/node/security_context.go:49 -[It] should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:346 -Dec 14 16:57:54.440: INFO: Waiting up to 5m0s for pod "busybox-user-65534-de44f4f5-749f-464d-a6d9-ab6accc504be" in namespace "security-context-test-510" to be "Succeeded or Failed" -Dec 14 16:57:54.443: INFO: Pod "busybox-user-65534-de44f4f5-749f-464d-a6d9-ab6accc504be": Phase="Pending", Reason="", readiness=false. Elapsed: 2.803664ms -Dec 14 16:57:56.450: INFO: Pod "busybox-user-65534-de44f4f5-749f-464d-a6d9-ab6accc504be": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010190002s -Dec 14 16:57:58.450: INFO: Pod "busybox-user-65534-de44f4f5-749f-464d-a6d9-ab6accc504be": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010174868s -Dec 14 16:57:58.450: INFO: Pod "busybox-user-65534-de44f4f5-749f-464d-a6d9-ab6accc504be" satisfied condition "Succeeded or Failed" -[AfterEach] [sig-node] Security Context +STEP: Creating a kubernetes client 02/25/23 11:28:10.694 +Feb 25 11:28:10.695: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 11:28:10.698 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:10.728 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:10.733 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 +STEP: Setting up server cert 02/25/23 11:28:10.762 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 11:28:11.417 +STEP: Deploying the webhook pod 02/25/23 11:28:11.429 +STEP: Wait for the deployment to be ready 02/25/23 11:28:11.455 +Feb 25 11:28:11.466: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created +STEP: Deploying the webhook service 02/25/23 11:28:13.501 +STEP: Verifying the service has paired with the endpoint 02/25/23 11:28:13.522 +Feb 25 11:28:14.522: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should mutate custom resource [Conformance] + test/e2e/apimachinery/webhook.go:290 +Feb 25 11:28:14.528: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Registering the mutating webhook for custom resource e2e-test-webhook-6297-crds.webhook.example.com via the AdmissionRegistration API 02/25/23 11:28:15.051 +STEP: Creating a custom resource that should be mutated by the webhook 02/25/23 11:28:15.087 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 16:57:58.450: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "security-context-test-510" for this suite. 12/14/22 16:57:58.458 -{"msg":"PASSED [sig-node] Security Context When creating a container with runAsUser should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance]","completed":300,"skipped":5709,"failed":0} +Feb 25 11:28:17.820: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-1775" for this suite. 02/25/23 11:28:17.827 +STEP: Destroying namespace "webhook-1775-markers" for this suite. 02/25/23 11:28:17.838 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate custom resource [Conformance]","completed":296,"skipped":5646,"failed":0} ------------------------------ -• [4.085 seconds] -[sig-node] Security Context -test/e2e/common/node/framework.go:23 - When creating a container with runAsUser - test/e2e/common/node/security_context.go:308 - should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:346 +• [SLOW TEST] [7.241 seconds] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + should mutate custom resource [Conformance] + test/e2e/apimachinery/webhook.go:290 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Security Context + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:57:54.388 - Dec 14 16:57:54.388: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename security-context-test 12/14/22 16:57:54.392 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:57:54.423 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:57:54.427 - [BeforeEach] [sig-node] Security Context - test/e2e/common/node/security_context.go:49 - [It] should run the container with uid 65534 [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:346 - Dec 14 16:57:54.440: INFO: Waiting up to 5m0s for pod "busybox-user-65534-de44f4f5-749f-464d-a6d9-ab6accc504be" in namespace "security-context-test-510" to be "Succeeded or Failed" - Dec 14 16:57:54.443: INFO: Pod "busybox-user-65534-de44f4f5-749f-464d-a6d9-ab6accc504be": Phase="Pending", Reason="", readiness=false. Elapsed: 2.803664ms - Dec 14 16:57:56.450: INFO: Pod "busybox-user-65534-de44f4f5-749f-464d-a6d9-ab6accc504be": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010190002s - Dec 14 16:57:58.450: INFO: Pod "busybox-user-65534-de44f4f5-749f-464d-a6d9-ab6accc504be": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.010174868s - Dec 14 16:57:58.450: INFO: Pod "busybox-user-65534-de44f4f5-749f-464d-a6d9-ab6accc504be" satisfied condition "Succeeded or Failed" - [AfterEach] [sig-node] Security Context + STEP: Creating a kubernetes client 02/25/23 11:28:10.694 + Feb 25 11:28:10.695: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 11:28:10.698 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:10.728 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:10.733 + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 + STEP: Setting up server cert 02/25/23 11:28:10.762 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 11:28:11.417 + STEP: Deploying the webhook pod 02/25/23 11:28:11.429 + STEP: Wait for the deployment to be ready 02/25/23 11:28:11.455 + Feb 25 11:28:11.466: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created + STEP: Deploying the webhook service 02/25/23 11:28:13.501 + STEP: Verifying the service has paired with the endpoint 02/25/23 11:28:13.522 + Feb 25 11:28:14.522: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] should mutate custom resource [Conformance] + test/e2e/apimachinery/webhook.go:290 + Feb 25 11:28:14.528: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Registering the mutating webhook for custom resource e2e-test-webhook-6297-crds.webhook.example.com via the AdmissionRegistration API 02/25/23 11:28:15.051 + STEP: Creating a custom resource that should be mutated by the webhook 02/25/23 11:28:15.087 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 16:57:58.450: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "security-context-test-510" for this suite. 12/14/22 16:57:58.458 + Feb 25 11:28:17.820: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-1775" for this suite. 02/25/23 11:28:17.827 + STEP: Destroying namespace "webhook-1775-markers" for this suite. 02/25/23 11:28:17.838 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-scheduling] SchedulerPreemption [Serial] - validates basic preemption works [Conformance] - test/e2e/scheduling/preemption.go:125 -[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] +[sig-apps] ReplicationController + should adopt matching pods on creation [Conformance] + test/e2e/apps/rc.go:91 +[BeforeEach] [sig-apps] ReplicationController test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:57:58.474 -Dec 14 16:57:58.474: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename sched-preemption 12/14/22 16:57:58.477 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:57:58.501 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:57:58.504 -[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:92 -Dec 14 16:57:58.539: INFO: Waiting up to 1m0s for all nodes to be ready -Dec 14 16:58:58.596: INFO: Waiting for terminating namespaces to be deleted... -[It] validates basic preemption works [Conformance] - test/e2e/scheduling/preemption.go:125 -STEP: Create pods that use 4/5 of node resources. 12/14/22 16:58:58.607 -Dec 14 16:58:58.684: INFO: Created pod: pod0-0-sched-preemption-low-priority -Dec 14 16:58:58.693: INFO: Created pod: pod0-1-sched-preemption-medium-priority -Dec 14 16:58:58.722: INFO: Created pod: pod1-0-sched-preemption-medium-priority -Dec 14 16:58:58.736: INFO: Created pod: pod1-1-sched-preemption-medium-priority -Dec 14 16:58:58.763: INFO: Created pod: pod2-0-sched-preemption-medium-priority -Dec 14 16:58:58.769: INFO: Created pod: pod2-1-sched-preemption-medium-priority -STEP: Wait for pods to be scheduled. 12/14/22 16:58:58.769 -Dec 14 16:58:58.770: INFO: Waiting up to 5m0s for pod "pod0-0-sched-preemption-low-priority" in namespace "sched-preemption-3222" to be "running" -Dec 14 16:58:58.775: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 5.267919ms -Dec 14 16:59:00.783: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012936122s -Dec 14 16:59:02.783: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 4.01278759s -Dec 14 16:59:04.783: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 6.013394072s -Dec 14 16:59:06.782: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 8.012051178s -Dec 14 16:59:08.781: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 10.011597016s -Dec 14 16:59:10.787: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Running", Reason="", readiness=true. Elapsed: 12.017671882s -Dec 14 16:59:10.788: INFO: Pod "pod0-0-sched-preemption-low-priority" satisfied condition "running" -Dec 14 16:59:10.788: INFO: Waiting up to 5m0s for pod "pod0-1-sched-preemption-medium-priority" in namespace "sched-preemption-3222" to be "running" -Dec 14 16:59:10.806: INFO: Pod "pod0-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 18.412493ms -Dec 14 16:59:10.806: INFO: Pod "pod0-1-sched-preemption-medium-priority" satisfied condition "running" -Dec 14 16:59:10.806: INFO: Waiting up to 5m0s for pod "pod1-0-sched-preemption-medium-priority" in namespace "sched-preemption-3222" to be "running" -Dec 14 16:59:10.809: INFO: Pod "pod1-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 3.002351ms -Dec 14 16:59:10.810: INFO: Pod "pod1-0-sched-preemption-medium-priority" satisfied condition "running" -Dec 14 16:59:10.810: INFO: Waiting up to 5m0s for pod "pod1-1-sched-preemption-medium-priority" in namespace "sched-preemption-3222" to be "running" -Dec 14 16:59:10.816: INFO: Pod "pod1-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 6.699904ms -Dec 14 16:59:10.816: INFO: Pod "pod1-1-sched-preemption-medium-priority" satisfied condition "running" -Dec 14 16:59:10.816: INFO: Waiting up to 5m0s for pod "pod2-0-sched-preemption-medium-priority" in namespace "sched-preemption-3222" to be "running" -Dec 14 16:59:10.820: INFO: Pod "pod2-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 3.461598ms -Dec 14 16:59:10.820: INFO: Pod "pod2-0-sched-preemption-medium-priority" satisfied condition "running" -Dec 14 16:59:10.820: INFO: Waiting up to 5m0s for pod "pod2-1-sched-preemption-medium-priority" in namespace "sched-preemption-3222" to be "running" -Dec 14 16:59:10.824: INFO: Pod "pod2-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 3.318183ms -Dec 14 16:59:10.824: INFO: Pod "pod2-1-sched-preemption-medium-priority" satisfied condition "running" -STEP: Run a high priority pod that has same requirements as that of lower priority pod 12/14/22 16:59:10.824 -Dec 14 16:59:10.831: INFO: Waiting up to 2m0s for pod "preemptor-pod" in namespace "sched-preemption-3222" to be "running" -Dec 14 16:59:10.835: INFO: Pod "preemptor-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 4.019572ms -Dec 14 16:59:12.843: INFO: Pod "preemptor-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011623435s -Dec 14 16:59:14.843: INFO: Pod "preemptor-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.011388285s -Dec 14 16:59:14.843: INFO: Pod "preemptor-pod" satisfied condition "running" -[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] +STEP: Creating a kubernetes client 02/25/23 11:28:17.966 +Feb 25 11:28:17.967: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename replication-controller 02/25/23 11:28:17.97 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:18.036 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:18.04 +[BeforeEach] [sig-apps] ReplicationController + test/e2e/apps/rc.go:56 +[It] should adopt matching pods on creation [Conformance] + test/e2e/apps/rc.go:91 +STEP: Given a Pod with a 'name' label pod-adoption is created 02/25/23 11:28:18.047 +Feb 25 11:28:18.075: INFO: Waiting up to 5m0s for pod "pod-adoption" in namespace "replication-controller-1702" to be "running and ready" +Feb 25 11:28:18.142: INFO: Pod "pod-adoption": Phase="Pending", Reason="", readiness=false. Elapsed: 67.311985ms +Feb 25 11:28:18.142: INFO: The phase of Pod pod-adoption is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:28:20.149: INFO: Pod "pod-adoption": Phase="Pending", Reason="", readiness=false. Elapsed: 2.073910017s +Feb 25 11:28:20.149: INFO: The phase of Pod pod-adoption is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:28:22.148: INFO: Pod "pod-adoption": Phase="Running", Reason="", readiness=true. Elapsed: 4.073145221s +Feb 25 11:28:22.148: INFO: The phase of Pod pod-adoption is Running (Ready = true) +Feb 25 11:28:22.148: INFO: Pod "pod-adoption" satisfied condition "running and ready" +STEP: When a replication controller with a matching selector is created 02/25/23 11:28:22.154 +STEP: Then the orphan pod is adopted 02/25/23 11:28:22.169 +[AfterEach] [sig-apps] ReplicationController test/e2e/framework/framework.go:187 -Dec 14 16:59:14.871: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "sched-preemption-3222" for this suite. 12/14/22 16:59:14.877 -[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:80 -{"msg":"PASSED [sig-scheduling] SchedulerPreemption [Serial] validates basic preemption works [Conformance]","completed":301,"skipped":5728,"failed":0} +Feb 25 11:28:23.185: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "replication-controller-1702" for this suite. 02/25/23 11:28:23.197 +{"msg":"PASSED [sig-apps] ReplicationController should adopt matching pods on creation [Conformance]","completed":297,"skipped":5667,"failed":0} ------------------------------ -• [SLOW TEST] [76.476 seconds] -[sig-scheduling] SchedulerPreemption [Serial] -test/e2e/scheduling/framework.go:40 - validates basic preemption works [Conformance] - test/e2e/scheduling/preemption.go:125 +• [SLOW TEST] [5.246 seconds] +[sig-apps] ReplicationController +test/e2e/apps/framework.go:23 + should adopt matching pods on creation [Conformance] + test/e2e/apps/rc.go:91 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] + [BeforeEach] [sig-apps] ReplicationController test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:57:58.474 - Dec 14 16:57:58.474: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename sched-preemption 12/14/22 16:57:58.477 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:57:58.501 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:57:58.504 - [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:92 - Dec 14 16:57:58.539: INFO: Waiting up to 1m0s for all nodes to be ready - Dec 14 16:58:58.596: INFO: Waiting for terminating namespaces to be deleted... - [It] validates basic preemption works [Conformance] - test/e2e/scheduling/preemption.go:125 - STEP: Create pods that use 4/5 of node resources. 12/14/22 16:58:58.607 - Dec 14 16:58:58.684: INFO: Created pod: pod0-0-sched-preemption-low-priority - Dec 14 16:58:58.693: INFO: Created pod: pod0-1-sched-preemption-medium-priority - Dec 14 16:58:58.722: INFO: Created pod: pod1-0-sched-preemption-medium-priority - Dec 14 16:58:58.736: INFO: Created pod: pod1-1-sched-preemption-medium-priority - Dec 14 16:58:58.763: INFO: Created pod: pod2-0-sched-preemption-medium-priority - Dec 14 16:58:58.769: INFO: Created pod: pod2-1-sched-preemption-medium-priority - STEP: Wait for pods to be scheduled. 12/14/22 16:58:58.769 - Dec 14 16:58:58.770: INFO: Waiting up to 5m0s for pod "pod0-0-sched-preemption-low-priority" in namespace "sched-preemption-3222" to be "running" - Dec 14 16:58:58.775: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 5.267919ms - Dec 14 16:59:00.783: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012936122s - Dec 14 16:59:02.783: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 4.01278759s - Dec 14 16:59:04.783: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 6.013394072s - Dec 14 16:59:06.782: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 8.012051178s - Dec 14 16:59:08.781: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 10.011597016s - Dec 14 16:59:10.787: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Running", Reason="", readiness=true. Elapsed: 12.017671882s - Dec 14 16:59:10.788: INFO: Pod "pod0-0-sched-preemption-low-priority" satisfied condition "running" - Dec 14 16:59:10.788: INFO: Waiting up to 5m0s for pod "pod0-1-sched-preemption-medium-priority" in namespace "sched-preemption-3222" to be "running" - Dec 14 16:59:10.806: INFO: Pod "pod0-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 18.412493ms - Dec 14 16:59:10.806: INFO: Pod "pod0-1-sched-preemption-medium-priority" satisfied condition "running" - Dec 14 16:59:10.806: INFO: Waiting up to 5m0s for pod "pod1-0-sched-preemption-medium-priority" in namespace "sched-preemption-3222" to be "running" - Dec 14 16:59:10.809: INFO: Pod "pod1-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 3.002351ms - Dec 14 16:59:10.810: INFO: Pod "pod1-0-sched-preemption-medium-priority" satisfied condition "running" - Dec 14 16:59:10.810: INFO: Waiting up to 5m0s for pod "pod1-1-sched-preemption-medium-priority" in namespace "sched-preemption-3222" to be "running" - Dec 14 16:59:10.816: INFO: Pod "pod1-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 6.699904ms - Dec 14 16:59:10.816: INFO: Pod "pod1-1-sched-preemption-medium-priority" satisfied condition "running" - Dec 14 16:59:10.816: INFO: Waiting up to 5m0s for pod "pod2-0-sched-preemption-medium-priority" in namespace "sched-preemption-3222" to be "running" - Dec 14 16:59:10.820: INFO: Pod "pod2-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 3.461598ms - Dec 14 16:59:10.820: INFO: Pod "pod2-0-sched-preemption-medium-priority" satisfied condition "running" - Dec 14 16:59:10.820: INFO: Waiting up to 5m0s for pod "pod2-1-sched-preemption-medium-priority" in namespace "sched-preemption-3222" to be "running" - Dec 14 16:59:10.824: INFO: Pod "pod2-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 3.318183ms - Dec 14 16:59:10.824: INFO: Pod "pod2-1-sched-preemption-medium-priority" satisfied condition "running" - STEP: Run a high priority pod that has same requirements as that of lower priority pod 12/14/22 16:59:10.824 - Dec 14 16:59:10.831: INFO: Waiting up to 2m0s for pod "preemptor-pod" in namespace "sched-preemption-3222" to be "running" - Dec 14 16:59:10.835: INFO: Pod "preemptor-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 4.019572ms - Dec 14 16:59:12.843: INFO: Pod "preemptor-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011623435s - Dec 14 16:59:14.843: INFO: Pod "preemptor-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.011388285s - Dec 14 16:59:14.843: INFO: Pod "preemptor-pod" satisfied condition "running" - [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] + STEP: Creating a kubernetes client 02/25/23 11:28:17.966 + Feb 25 11:28:17.967: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename replication-controller 02/25/23 11:28:17.97 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:18.036 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:18.04 + [BeforeEach] [sig-apps] ReplicationController + test/e2e/apps/rc.go:56 + [It] should adopt matching pods on creation [Conformance] + test/e2e/apps/rc.go:91 + STEP: Given a Pod with a 'name' label pod-adoption is created 02/25/23 11:28:18.047 + Feb 25 11:28:18.075: INFO: Waiting up to 5m0s for pod "pod-adoption" in namespace "replication-controller-1702" to be "running and ready" + Feb 25 11:28:18.142: INFO: Pod "pod-adoption": Phase="Pending", Reason="", readiness=false. Elapsed: 67.311985ms + Feb 25 11:28:18.142: INFO: The phase of Pod pod-adoption is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:28:20.149: INFO: Pod "pod-adoption": Phase="Pending", Reason="", readiness=false. Elapsed: 2.073910017s + Feb 25 11:28:20.149: INFO: The phase of Pod pod-adoption is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:28:22.148: INFO: Pod "pod-adoption": Phase="Running", Reason="", readiness=true. Elapsed: 4.073145221s + Feb 25 11:28:22.148: INFO: The phase of Pod pod-adoption is Running (Ready = true) + Feb 25 11:28:22.148: INFO: Pod "pod-adoption" satisfied condition "running and ready" + STEP: When a replication controller with a matching selector is created 02/25/23 11:28:22.154 + STEP: Then the orphan pod is adopted 02/25/23 11:28:22.169 + [AfterEach] [sig-apps] ReplicationController test/e2e/framework/framework.go:187 - Dec 14 16:59:14.871: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "sched-preemption-3222" for this suite. 12/14/22 16:59:14.877 - [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] - test/e2e/scheduling/preemption.go:80 + Feb 25 11:28:23.185: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "replication-controller-1702" for this suite. 02/25/23 11:28:23.197 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSSSSSS ------------------------------ -[sig-api-machinery] Aggregator - Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] - test/e2e/apimachinery/aggregator.go:100 -[BeforeEach] [sig-api-machinery] Aggregator +[sig-apps] DisruptionController + should block an eviction until the PDB is updated to allow it [Conformance] + test/e2e/apps/disruption.go:346 +[BeforeEach] [sig-apps] DisruptionController test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:59:14.954 -Dec 14 16:59:14.954: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename aggregator 12/14/22 16:59:14.956 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:59:14.985 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:59:14.989 -[BeforeEach] [sig-api-machinery] Aggregator - test/e2e/apimachinery/aggregator.go:78 -Dec 14 16:59:14.992: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -[It] Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] - test/e2e/apimachinery/aggregator.go:100 -STEP: Registering the sample API server. 12/14/22 16:59:14.993 -Dec 14 16:59:16.348: INFO: new replicaset for deployment "sample-apiserver-deployment" is yet to be created -Dec 14 16:59:18.410: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:59:20.415: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:59:22.419: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:59:24.430: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:59:26.419: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:59:28.421: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:59:30.415: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:59:32.420: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:59:34.419: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:59:36.421: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:59:38.420: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:59:40.416: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:59:42.421: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 16:59:44.585: INFO: Waited 157.379632ms for the sample-apiserver to be ready to handle requests. -STEP: Read Status for v1alpha1.wardle.example.com 12/14/22 16:59:44.704 -STEP: kubectl patch apiservice v1alpha1.wardle.example.com -p '{"spec":{"versionPriority": 400}}' 12/14/22 16:59:44.709 -STEP: List APIServices 12/14/22 16:59:44.72 -Dec 14 16:59:44.738: INFO: Found v1alpha1.wardle.example.com in APIServiceList -[AfterEach] [sig-api-machinery] Aggregator - test/e2e/apimachinery/aggregator.go:68 -[AfterEach] [sig-api-machinery] Aggregator +STEP: Creating a kubernetes client 02/25/23 11:28:23.223 +Feb 25 11:28:23.223: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename disruption 02/25/23 11:28:23.225 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:23.272 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:23.277 +[BeforeEach] [sig-apps] DisruptionController + test/e2e/apps/disruption.go:71 +[It] should block an eviction until the PDB is updated to allow it [Conformance] + test/e2e/apps/disruption.go:346 +STEP: Creating a pdb that targets all three pods in a test replica set 02/25/23 11:28:23.281 +STEP: Waiting for the pdb to be processed 02/25/23 11:28:23.291 +STEP: First trying to evict a pod which shouldn't be evictable 02/25/23 11:28:25.325 +STEP: Waiting for all pods to be running 02/25/23 11:28:25.325 +Feb 25 11:28:25.334: INFO: pods: 0 < 3 +Feb 25 11:28:27.343: INFO: running pods: 2 < 3 +STEP: locating a running pod 02/25/23 11:28:29.341 +STEP: Updating the pdb to allow a pod to be evicted 02/25/23 11:28:29.37 +STEP: Waiting for the pdb to be processed 02/25/23 11:28:29.387 +STEP: Trying to evict the same pod we tried earlier which should now be evictable 02/25/23 11:28:31.414 +STEP: Waiting for all pods to be running 02/25/23 11:28:31.415 +STEP: Waiting for the pdb to observed all healthy pods 02/25/23 11:28:31.426 +STEP: Patching the pdb to disallow a pod to be evicted 02/25/23 11:28:31.466 +STEP: Waiting for the pdb to be processed 02/25/23 11:28:31.512 +STEP: Waiting for all pods to be running 02/25/23 11:28:33.525 +STEP: locating a running pod 02/25/23 11:28:33.53 +STEP: Deleting the pdb to allow a pod to be evicted 02/25/23 11:28:33.549 +STEP: Waiting for the pdb to be deleted 02/25/23 11:28:33.558 +STEP: Trying to evict the same pod we tried earlier which should now be evictable 02/25/23 11:28:33.562 +STEP: Waiting for all pods to be running 02/25/23 11:28:33.563 +[AfterEach] [sig-apps] DisruptionController test/e2e/framework/framework.go:187 -Dec 14 16:59:44.916: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "aggregator-8562" for this suite. 12/14/22 16:59:44.925 -{"msg":"PASSED [sig-api-machinery] Aggregator Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance]","completed":302,"skipped":5731,"failed":0} +Feb 25 11:28:33.585: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "disruption-1218" for this suite. 02/25/23 11:28:33.593 +{"msg":"PASSED [sig-apps] DisruptionController should block an eviction until the PDB is updated to allow it [Conformance]","completed":298,"skipped":5677,"failed":0} ------------------------------ -• [SLOW TEST] [29.983 seconds] -[sig-api-machinery] Aggregator -test/e2e/apimachinery/framework.go:23 - Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] - test/e2e/apimachinery/aggregator.go:100 +• [SLOW TEST] [10.437 seconds] +[sig-apps] DisruptionController +test/e2e/apps/framework.go:23 + should block an eviction until the PDB is updated to allow it [Conformance] + test/e2e/apps/disruption.go:346 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Aggregator + [BeforeEach] [sig-apps] DisruptionController test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:59:14.954 - Dec 14 16:59:14.954: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename aggregator 12/14/22 16:59:14.956 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:59:14.985 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:59:14.989 - [BeforeEach] [sig-api-machinery] Aggregator - test/e2e/apimachinery/aggregator.go:78 - Dec 14 16:59:14.992: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - [It] Should be able to support the 1.17 Sample API Server using the current Aggregator [Conformance] - test/e2e/apimachinery/aggregator.go:100 - STEP: Registering the sample API server. 12/14/22 16:59:14.993 - Dec 14 16:59:16.348: INFO: new replicaset for deployment "sample-apiserver-deployment" is yet to be created - Dec 14 16:59:18.410: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:59:20.415: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:59:22.419: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:59:24.430: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:59:26.419: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:59:28.421: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:59:30.415: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:59:32.420: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:59:34.419: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:59:36.421: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:59:38.420: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:59:40.416: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:59:42.421: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 16, 59, 16, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-apiserver-deployment-5885c99c55\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 16:59:44.585: INFO: Waited 157.379632ms for the sample-apiserver to be ready to handle requests. - STEP: Read Status for v1alpha1.wardle.example.com 12/14/22 16:59:44.704 - STEP: kubectl patch apiservice v1alpha1.wardle.example.com -p '{"spec":{"versionPriority": 400}}' 12/14/22 16:59:44.709 - STEP: List APIServices 12/14/22 16:59:44.72 - Dec 14 16:59:44.738: INFO: Found v1alpha1.wardle.example.com in APIServiceList - [AfterEach] [sig-api-machinery] Aggregator - test/e2e/apimachinery/aggregator.go:68 - [AfterEach] [sig-api-machinery] Aggregator + STEP: Creating a kubernetes client 02/25/23 11:28:23.223 + Feb 25 11:28:23.223: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename disruption 02/25/23 11:28:23.225 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:23.272 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:23.277 + [BeforeEach] [sig-apps] DisruptionController + test/e2e/apps/disruption.go:71 + [It] should block an eviction until the PDB is updated to allow it [Conformance] + test/e2e/apps/disruption.go:346 + STEP: Creating a pdb that targets all three pods in a test replica set 02/25/23 11:28:23.281 + STEP: Waiting for the pdb to be processed 02/25/23 11:28:23.291 + STEP: First trying to evict a pod which shouldn't be evictable 02/25/23 11:28:25.325 + STEP: Waiting for all pods to be running 02/25/23 11:28:25.325 + Feb 25 11:28:25.334: INFO: pods: 0 < 3 + Feb 25 11:28:27.343: INFO: running pods: 2 < 3 + STEP: locating a running pod 02/25/23 11:28:29.341 + STEP: Updating the pdb to allow a pod to be evicted 02/25/23 11:28:29.37 + STEP: Waiting for the pdb to be processed 02/25/23 11:28:29.387 + STEP: Trying to evict the same pod we tried earlier which should now be evictable 02/25/23 11:28:31.414 + STEP: Waiting for all pods to be running 02/25/23 11:28:31.415 + STEP: Waiting for the pdb to observed all healthy pods 02/25/23 11:28:31.426 + STEP: Patching the pdb to disallow a pod to be evicted 02/25/23 11:28:31.466 + STEP: Waiting for the pdb to be processed 02/25/23 11:28:31.512 + STEP: Waiting for all pods to be running 02/25/23 11:28:33.525 + STEP: locating a running pod 02/25/23 11:28:33.53 + STEP: Deleting the pdb to allow a pod to be evicted 02/25/23 11:28:33.549 + STEP: Waiting for the pdb to be deleted 02/25/23 11:28:33.558 + STEP: Trying to evict the same pod we tried earlier which should now be evictable 02/25/23 11:28:33.562 + STEP: Waiting for all pods to be running 02/25/23 11:28:33.563 + [AfterEach] [sig-apps] DisruptionController test/e2e/framework/framework.go:187 - Dec 14 16:59:44.916: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "aggregator-8562" for this suite. 12/14/22 16:59:44.925 + Feb 25 11:28:33.585: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "disruption-1218" for this suite. 02/25/23 11:28:33.593 << End Captured GinkgoWriter Output ------------------------------ -SSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-network] EndpointSlice - should support creating EndpointSlice API operations [Conformance] - test/e2e/network/endpointslice.go:352 -[BeforeEach] [sig-network] EndpointSlice +[sig-node] Pods + should patch a pod status [Conformance] + test/e2e/common/node/pods.go:1082 +[BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:59:44.938 -Dec 14 16:59:44.938: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename endpointslice 12/14/22 16:59:44.94 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:59:45.009 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:59:45.013 -[BeforeEach] [sig-network] EndpointSlice - test/e2e/network/endpointslice.go:51 -[It] should support creating EndpointSlice API operations [Conformance] - test/e2e/network/endpointslice.go:352 -STEP: getting /apis 12/14/22 16:59:45.019 -STEP: getting /apis/discovery.k8s.io 12/14/22 16:59:45.023 -STEP: getting /apis/discovery.k8s.iov1 12/14/22 16:59:45.028 -STEP: creating 12/14/22 16:59:45.03 -STEP: getting 12/14/22 16:59:45.057 -STEP: listing 12/14/22 16:59:45.061 -STEP: watching 12/14/22 16:59:45.064 -Dec 14 16:59:45.064: INFO: starting watch -STEP: cluster-wide listing 12/14/22 16:59:45.066 -STEP: cluster-wide watching 12/14/22 16:59:45.069 -Dec 14 16:59:45.069: INFO: starting watch -STEP: patching 12/14/22 16:59:45.07 -STEP: updating 12/14/22 16:59:45.176 -Dec 14 16:59:45.187: INFO: waiting for watch events with expected annotations -Dec 14 16:59:45.187: INFO: saw patched and updated annotations -STEP: deleting 12/14/22 16:59:45.187 -STEP: deleting a collection 12/14/22 16:59:45.235 -[AfterEach] [sig-network] EndpointSlice +STEP: Creating a kubernetes client 02/25/23 11:28:33.67 +Feb 25 11:28:33.670: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename pods 02/25/23 11:28:33.679 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:33.719 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:33.724 +[BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 +[It] should patch a pod status [Conformance] + test/e2e/common/node/pods.go:1082 +STEP: Create a pod 02/25/23 11:28:33.728 +Feb 25 11:28:33.749: INFO: Waiting up to 5m0s for pod "pod-cvvpr" in namespace "pods-867" to be "running" +Feb 25 11:28:33.755: INFO: Pod "pod-cvvpr": Phase="Pending", Reason="", readiness=false. Elapsed: 6.118472ms +Feb 25 11:28:35.765: INFO: Pod "pod-cvvpr": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015844369s +Feb 25 11:28:37.763: INFO: Pod "pod-cvvpr": Phase="Running", Reason="", readiness=true. Elapsed: 4.014485388s +Feb 25 11:28:37.763: INFO: Pod "pod-cvvpr" satisfied condition "running" +STEP: patching /status 02/25/23 11:28:37.764 +Feb 25 11:28:37.774: INFO: Status Message: "Patched by e2e test" and Reason: "E2E" +[AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 -Dec 14 16:59:45.285: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "endpointslice-3542" for this suite. 12/14/22 16:59:45.292 -{"msg":"PASSED [sig-network] EndpointSlice should support creating EndpointSlice API operations [Conformance]","completed":303,"skipped":5737,"failed":0} +Feb 25 11:28:37.775: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-867" for this suite. 02/25/23 11:28:37.782 +{"msg":"PASSED [sig-node] Pods should patch a pod status [Conformance]","completed":299,"skipped":5733,"failed":0} ------------------------------ -• [0.369 seconds] -[sig-network] EndpointSlice -test/e2e/network/common/framework.go:23 - should support creating EndpointSlice API operations [Conformance] - test/e2e/network/endpointslice.go:352 +• [4.126 seconds] +[sig-node] Pods +test/e2e/common/node/framework.go:23 + should patch a pod status [Conformance] + test/e2e/common/node/pods.go:1082 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] EndpointSlice + [BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:59:44.938 - Dec 14 16:59:44.938: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename endpointslice 12/14/22 16:59:44.94 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:59:45.009 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:59:45.013 - [BeforeEach] [sig-network] EndpointSlice - test/e2e/network/endpointslice.go:51 - [It] should support creating EndpointSlice API operations [Conformance] - test/e2e/network/endpointslice.go:352 - STEP: getting /apis 12/14/22 16:59:45.019 - STEP: getting /apis/discovery.k8s.io 12/14/22 16:59:45.023 - STEP: getting /apis/discovery.k8s.iov1 12/14/22 16:59:45.028 - STEP: creating 12/14/22 16:59:45.03 - STEP: getting 12/14/22 16:59:45.057 - STEP: listing 12/14/22 16:59:45.061 - STEP: watching 12/14/22 16:59:45.064 - Dec 14 16:59:45.064: INFO: starting watch - STEP: cluster-wide listing 12/14/22 16:59:45.066 - STEP: cluster-wide watching 12/14/22 16:59:45.069 - Dec 14 16:59:45.069: INFO: starting watch - STEP: patching 12/14/22 16:59:45.07 - STEP: updating 12/14/22 16:59:45.176 - Dec 14 16:59:45.187: INFO: waiting for watch events with expected annotations - Dec 14 16:59:45.187: INFO: saw patched and updated annotations - STEP: deleting 12/14/22 16:59:45.187 - STEP: deleting a collection 12/14/22 16:59:45.235 - [AfterEach] [sig-network] EndpointSlice + STEP: Creating a kubernetes client 02/25/23 11:28:33.67 + Feb 25 11:28:33.670: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename pods 02/25/23 11:28:33.679 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:33.719 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:33.724 + [BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 + [It] should patch a pod status [Conformance] + test/e2e/common/node/pods.go:1082 + STEP: Create a pod 02/25/23 11:28:33.728 + Feb 25 11:28:33.749: INFO: Waiting up to 5m0s for pod "pod-cvvpr" in namespace "pods-867" to be "running" + Feb 25 11:28:33.755: INFO: Pod "pod-cvvpr": Phase="Pending", Reason="", readiness=false. Elapsed: 6.118472ms + Feb 25 11:28:35.765: INFO: Pod "pod-cvvpr": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015844369s + Feb 25 11:28:37.763: INFO: Pod "pod-cvvpr": Phase="Running", Reason="", readiness=true. Elapsed: 4.014485388s + Feb 25 11:28:37.763: INFO: Pod "pod-cvvpr" satisfied condition "running" + STEP: patching /status 02/25/23 11:28:37.764 + Feb 25 11:28:37.774: INFO: Status Message: "Patched by e2e test" and Reason: "E2E" + [AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 - Dec 14 16:59:45.285: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "endpointslice-3542" for this suite. 12/14/22 16:59:45.292 + Feb 25 11:28:37.775: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "pods-867" for this suite. 02/25/23 11:28:37.782 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-network] EndpointSlice - should create and delete Endpoints and EndpointSlices for a Service with a selector specified [Conformance] - test/e2e/network/endpointslice.go:101 -[BeforeEach] [sig-network] EndpointSlice +[sig-apps] Daemon set [Serial] + should list and delete a collection of DaemonSets [Conformance] + test/e2e/apps/daemon_set.go:822 +[BeforeEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:59:45.31 -Dec 14 16:59:45.310: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename endpointslice 12/14/22 16:59:45.313 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:59:45.367 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:59:45.37 -[BeforeEach] [sig-network] EndpointSlice - test/e2e/network/endpointslice.go:51 -[It] should create and delete Endpoints and EndpointSlices for a Service with a selector specified [Conformance] - test/e2e/network/endpointslice.go:101 -[AfterEach] [sig-network] EndpointSlice +STEP: Creating a kubernetes client 02/25/23 11:28:37.806 +Feb 25 11:28:37.806: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename daemonsets 02/25/23 11:28:37.809 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:37.832 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:37.835 +[BeforeEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:145 +[It] should list and delete a collection of DaemonSets [Conformance] + test/e2e/apps/daemon_set.go:822 +STEP: Creating simple DaemonSet "daemon-set" 02/25/23 11:28:37.872 +STEP: Check that daemon pods launch on every node of the cluster. 02/25/23 11:28:37.881 +Feb 25 11:28:37.894: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 11:28:37.894: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 11:28:38.942: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 +Feb 25 11:28:38.942: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 11:28:39.922: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 +Feb 25 11:28:39.922: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 +Feb 25 11:28:40.964: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 +Feb 25 11:28:40.964: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set +STEP: listing all DeamonSets 02/25/23 11:28:40.969 +STEP: DeleteCollection of the DaemonSets 02/25/23 11:28:40.992 +STEP: Verify that ReplicaSets have been deleted 02/25/23 11:28:41.007 +[AfterEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:110 +Feb 25 11:28:41.077: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"42804"},"items":null} + +Feb 25 11:28:41.094: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"42806"},"items":[{"metadata":{"name":"daemon-set-ch42b","generateName":"daemon-set-","namespace":"daemonsets-8281","uid":"00df658b-5259-4943-a1cd-2849001ab730","resourceVersion":"42803","creationTimestamp":"2023-02-25T11:28:37Z","deletionTimestamp":"2023-02-25T11:29:11Z","deletionGracePeriodSeconds":30,"labels":{"controller-revision-hash":"7f7ffb4fcc","daemonset-name":"daemon-set","pod-template-generation":"1"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"DaemonSet","name":"daemon-set","uid":"ba66bb7e-57bb-4b5c-9258-901aa735d014","controller":true,"blockOwnerDeletion":true}],"managedFields":[{"manager":"kube-controller-manager","operation":"Update","apiVersion":"v1","time":"2023-02-25T11:28:37Z","fieldsType":"FieldsV1","fieldsV1":{"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:controller-revision-hash":{},"f:daemonset-name":{},"f:pod-template-generation":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"ba66bb7e-57bb-4b5c-9258-901aa735d014\"}":{}}},"f:spec":{"f:affinity":{".":{},"f:nodeAffinity":{".":{},"f:requiredDuringSchedulingIgnoredDuringExecution":{}}},"f:containers":{"k:{\"name\":\"app\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:ports":{".":{},"k:{\"containerPort\":9376,\"protocol\":\"TCP\"}":{".":{},"f:containerPort":{},"f:protocol":{}}},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{},"f:tolerations":{}}}},{"manager":"kubelet","operation":"Update","apiVersion":"v1","time":"2023-02-25T11:28:40Z","fieldsType":"FieldsV1","fieldsV1":{"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.8\"}":{".":{},"f:ip":{}}},"f:startTime":{}}},"subresource":"status"}]},"spec":{"volumes":[{"name":"kube-api-access-hr767","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"app","image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","ports":[{"containerPort":9376,"protocol":"TCP"}],"resources":{},"volumeMounts":[{"name":"kube-api-access-hr767","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"File","imagePullPolicy":"IfNotPresent","securityContext":{}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","serviceAccountName":"default","serviceAccount":"default","nodeName":"diek3dua9gay-3","securityContext":{},"affinity":{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchFields":[{"key":"metadata.name","operator":"In","values":["diek3dua9gay-3"]}]}]}}},"schedulerName":"default-scheduler","tolerations":[{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/disk-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/pid-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unschedulable","operator":"Exists","effect":"NoSchedule"}],"priority":0,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:38Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:40Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:40Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:37Z"}],"hostIP":"192.168.121.44","podIP":"10.233.66.8","podIPs":[{"ip":"10.233.66.8"}],"startTime":"2023-02-25T11:28:38Z","containerStatuses":[{"name":"app","state":{"running":{"startedAt":"2023-02-25T11:28:40Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","imageID":"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3","containerID":"cri-o://6fe0b31aa809189aed1096353c47cfa992a03a1142224390f1821fbd68b26430","started":true}],"qosClass":"BestEffort"}},{"metadata":{"name":"daemon-set-h5542","generateName":"daemon-set-","namespace":"daemonsets-8281","uid":"e97d4e46-4931-4418-afee-e1c94445f153","resourceVersion":"42806","creationTimestamp":"2023-02-25T11:28:37Z","deletionTimestamp":"2023-02-25T11:29:11Z","deletionGracePeriodSeconds":30,"labels":{"controller-revision-hash":"7f7ffb4fcc","daemonset-name":"daemon-set","pod-template-generation":"1"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"DaemonSet","name":"daemon-set","uid":"ba66bb7e-57bb-4b5c-9258-901aa735d014","controller":true,"blockOwnerDeletion":true}],"managedFields":[{"manager":"kube-controller-manager","operation":"Update","apiVersion":"v1","time":"2023-02-25T11:28:37Z","fieldsType":"FieldsV1","fieldsV1":{"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:controller-revision-hash":{},"f:daemonset-name":{},"f:pod-template-generation":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"ba66bb7e-57bb-4b5c-9258-901aa735d014\"}":{}}},"f:spec":{"f:affinity":{".":{},"f:nodeAffinity":{".":{},"f:requiredDuringSchedulingIgnoredDuringExecution":{}}},"f:containers":{"k:{\"name\":\"app\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:ports":{".":{},"k:{\"containerPort\":9376,\"protocol\":\"TCP\"}":{".":{},"f:containerPort":{},"f:protocol":{}}},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{},"f:tolerations":{}}}},{"manager":"kubelet","operation":"Update","apiVersion":"v1","time":"2023-02-25T11:28:39Z","fieldsType":"FieldsV1","fieldsV1":{"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.223\"}":{".":{},"f:ip":{}}},"f:startTime":{}}},"subresource":"status"}]},"spec":{"volumes":[{"name":"kube-api-access-p7qkn","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"app","image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","ports":[{"containerPort":9376,"protocol":"TCP"}],"resources":{},"volumeMounts":[{"name":"kube-api-access-p7qkn","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"File","imagePullPolicy":"IfNotPresent","securityContext":{}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","serviceAccountName":"default","serviceAccount":"default","nodeName":"diek3dua9gay-2","securityContext":{},"affinity":{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchFields":[{"key":"metadata.name","operator":"In","values":["diek3dua9gay-2"]}]}]}}},"schedulerName":"default-scheduler","tolerations":[{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/disk-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/pid-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unschedulable","operator":"Exists","effect":"NoSchedule"}],"priority":0,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:37Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:39Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:39Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:37Z"}],"hostIP":"192.168.121.117","podIP":"10.233.64.223","podIPs":[{"ip":"10.233.64.223"}],"startTime":"2023-02-25T11:28:37Z","containerStatuses":[{"name":"app","state":{"running":{"startedAt":"2023-02-25T11:28:38Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","imageID":"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3","containerID":"cri-o://6024590c52b6d7cc0372abb02e11c0c3eb487e06eac17c7559696d9c121d2c96","started":true}],"qosClass":"BestEffort"}},{"metadata":{"name":"daemon-set-k7fsc","generateName":"daemon-set-","namespace":"daemonsets-8281","uid":"5fff7f57-4a9f-427f-8aa7-9d747ac7487c","resourceVersion":"42805","creationTimestamp":"2023-02-25T11:28:37Z","deletionTimestamp":"2023-02-25T11:29:11Z","deletionGracePeriodSeconds":30,"labels":{"controller-revision-hash":"7f7ffb4fcc","daemonset-name":"daemon-set","pod-template-generation":"1"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"DaemonSet","name":"daemon-set","uid":"ba66bb7e-57bb-4b5c-9258-901aa735d014","controller":true,"blockOwnerDeletion":true}],"managedFields":[{"manager":"kube-controller-manager","operation":"Update","apiVersion":"v1","time":"2023-02-25T11:28:37Z","fieldsType":"FieldsV1","fieldsV1":{"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:controller-revision-hash":{},"f:daemonset-name":{},"f:pod-template-generation":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"ba66bb7e-57bb-4b5c-9258-901aa735d014\"}":{}}},"f:spec":{"f:affinity":{".":{},"f:nodeAffinity":{".":{},"f:requiredDuringSchedulingIgnoredDuringExecution":{}}},"f:containers":{"k:{\"name\":\"app\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:ports":{".":{},"k:{\"containerPort\":9376,\"protocol\":\"TCP\"}":{".":{},"f:containerPort":{},"f:protocol":{}}},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{},"f:tolerations":{}}}},{"manager":"kubelet","operation":"Update","apiVersion":"v1","time":"2023-02-25T11:28:40Z","fieldsType":"FieldsV1","fieldsV1":{"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.141\"}":{".":{},"f:ip":{}}},"f:startTime":{}}},"subresource":"status"}]},"spec":{"volumes":[{"name":"kube-api-access-j5mdp","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"app","image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","ports":[{"containerPort":9376,"protocol":"TCP"}],"resources":{},"volumeMounts":[{"name":"kube-api-access-j5mdp","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"File","imagePullPolicy":"IfNotPresent","securityContext":{}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","serviceAccountName":"default","serviceAccount":"default","nodeName":"diek3dua9gay-1","securityContext":{},"affinity":{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchFields":[{"key":"metadata.name","operator":"In","values":["diek3dua9gay-1"]}]}]}}},"schedulerName":"default-scheduler","tolerations":[{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/disk-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/pid-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unschedulable","operator":"Exists","effect":"NoSchedule"}],"priority":0,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:37Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:40Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:40Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:37Z"}],"hostIP":"192.168.121.250","podIP":"10.233.65.141","podIPs":[{"ip":"10.233.65.141"}],"startTime":"2023-02-25T11:28:37Z","containerStatuses":[{"name":"app","state":{"running":{"startedAt":"2023-02-25T11:28:39Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","imageID":"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3","containerID":"cri-o://20b693ee8e48492d54ffb578dd372904c7a2341a34c2f3d1a883e71bad1ce2be","started":true}],"qosClass":"BestEffort"}}]} + +[AfterEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:187 -Dec 14 16:59:47.462: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "endpointslice-9032" for this suite. 12/14/22 16:59:47.47 -{"msg":"PASSED [sig-network] EndpointSlice should create and delete Endpoints and EndpointSlices for a Service with a selector specified [Conformance]","completed":304,"skipped":5783,"failed":0} +Feb 25 11:28:41.261: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "daemonsets-8281" for this suite. 02/25/23 11:28:41.273 +{"msg":"PASSED [sig-apps] Daemon set [Serial] should list and delete a collection of DaemonSets [Conformance]","completed":300,"skipped":5766,"failed":0} ------------------------------ -• [2.169 seconds] -[sig-network] EndpointSlice -test/e2e/network/common/framework.go:23 - should create and delete Endpoints and EndpointSlices for a Service with a selector specified [Conformance] - test/e2e/network/endpointslice.go:101 +• [3.487 seconds] +[sig-apps] Daemon set [Serial] +test/e2e/apps/framework.go:23 + should list and delete a collection of DaemonSets [Conformance] + test/e2e/apps/daemon_set.go:822 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] EndpointSlice + [BeforeEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:59:45.31 - Dec 14 16:59:45.310: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename endpointslice 12/14/22 16:59:45.313 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:59:45.367 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:59:45.37 - [BeforeEach] [sig-network] EndpointSlice - test/e2e/network/endpointslice.go:51 - [It] should create and delete Endpoints and EndpointSlices for a Service with a selector specified [Conformance] - test/e2e/network/endpointslice.go:101 - [AfterEach] [sig-network] EndpointSlice + STEP: Creating a kubernetes client 02/25/23 11:28:37.806 + Feb 25 11:28:37.806: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename daemonsets 02/25/23 11:28:37.809 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:37.832 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:37.835 + [BeforeEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:145 + [It] should list and delete a collection of DaemonSets [Conformance] + test/e2e/apps/daemon_set.go:822 + STEP: Creating simple DaemonSet "daemon-set" 02/25/23 11:28:37.872 + STEP: Check that daemon pods launch on every node of the cluster. 02/25/23 11:28:37.881 + Feb 25 11:28:37.894: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 11:28:37.894: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 11:28:38.942: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 + Feb 25 11:28:38.942: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 11:28:39.922: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 + Feb 25 11:28:39.922: INFO: Node diek3dua9gay-1 is running 0 daemon pod, expected 1 + Feb 25 11:28:40.964: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 + Feb 25 11:28:40.964: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set + STEP: listing all DeamonSets 02/25/23 11:28:40.969 + STEP: DeleteCollection of the DaemonSets 02/25/23 11:28:40.992 + STEP: Verify that ReplicaSets have been deleted 02/25/23 11:28:41.007 + [AfterEach] [sig-apps] Daemon set [Serial] + test/e2e/apps/daemon_set.go:110 + Feb 25 11:28:41.077: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"42804"},"items":null} + + Feb 25 11:28:41.094: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"42806"},"items":[{"metadata":{"name":"daemon-set-ch42b","generateName":"daemon-set-","namespace":"daemonsets-8281","uid":"00df658b-5259-4943-a1cd-2849001ab730","resourceVersion":"42803","creationTimestamp":"2023-02-25T11:28:37Z","deletionTimestamp":"2023-02-25T11:29:11Z","deletionGracePeriodSeconds":30,"labels":{"controller-revision-hash":"7f7ffb4fcc","daemonset-name":"daemon-set","pod-template-generation":"1"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"DaemonSet","name":"daemon-set","uid":"ba66bb7e-57bb-4b5c-9258-901aa735d014","controller":true,"blockOwnerDeletion":true}],"managedFields":[{"manager":"kube-controller-manager","operation":"Update","apiVersion":"v1","time":"2023-02-25T11:28:37Z","fieldsType":"FieldsV1","fieldsV1":{"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:controller-revision-hash":{},"f:daemonset-name":{},"f:pod-template-generation":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"ba66bb7e-57bb-4b5c-9258-901aa735d014\"}":{}}},"f:spec":{"f:affinity":{".":{},"f:nodeAffinity":{".":{},"f:requiredDuringSchedulingIgnoredDuringExecution":{}}},"f:containers":{"k:{\"name\":\"app\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:ports":{".":{},"k:{\"containerPort\":9376,\"protocol\":\"TCP\"}":{".":{},"f:containerPort":{},"f:protocol":{}}},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{},"f:tolerations":{}}}},{"manager":"kubelet","operation":"Update","apiVersion":"v1","time":"2023-02-25T11:28:40Z","fieldsType":"FieldsV1","fieldsV1":{"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.8\"}":{".":{},"f:ip":{}}},"f:startTime":{}}},"subresource":"status"}]},"spec":{"volumes":[{"name":"kube-api-access-hr767","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"app","image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","ports":[{"containerPort":9376,"protocol":"TCP"}],"resources":{},"volumeMounts":[{"name":"kube-api-access-hr767","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"File","imagePullPolicy":"IfNotPresent","securityContext":{}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","serviceAccountName":"default","serviceAccount":"default","nodeName":"diek3dua9gay-3","securityContext":{},"affinity":{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchFields":[{"key":"metadata.name","operator":"In","values":["diek3dua9gay-3"]}]}]}}},"schedulerName":"default-scheduler","tolerations":[{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/disk-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/pid-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unschedulable","operator":"Exists","effect":"NoSchedule"}],"priority":0,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:38Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:40Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:40Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:37Z"}],"hostIP":"192.168.121.44","podIP":"10.233.66.8","podIPs":[{"ip":"10.233.66.8"}],"startTime":"2023-02-25T11:28:38Z","containerStatuses":[{"name":"app","state":{"running":{"startedAt":"2023-02-25T11:28:40Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","imageID":"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3","containerID":"cri-o://6fe0b31aa809189aed1096353c47cfa992a03a1142224390f1821fbd68b26430","started":true}],"qosClass":"BestEffort"}},{"metadata":{"name":"daemon-set-h5542","generateName":"daemon-set-","namespace":"daemonsets-8281","uid":"e97d4e46-4931-4418-afee-e1c94445f153","resourceVersion":"42806","creationTimestamp":"2023-02-25T11:28:37Z","deletionTimestamp":"2023-02-25T11:29:11Z","deletionGracePeriodSeconds":30,"labels":{"controller-revision-hash":"7f7ffb4fcc","daemonset-name":"daemon-set","pod-template-generation":"1"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"DaemonSet","name":"daemon-set","uid":"ba66bb7e-57bb-4b5c-9258-901aa735d014","controller":true,"blockOwnerDeletion":true}],"managedFields":[{"manager":"kube-controller-manager","operation":"Update","apiVersion":"v1","time":"2023-02-25T11:28:37Z","fieldsType":"FieldsV1","fieldsV1":{"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:controller-revision-hash":{},"f:daemonset-name":{},"f:pod-template-generation":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"ba66bb7e-57bb-4b5c-9258-901aa735d014\"}":{}}},"f:spec":{"f:affinity":{".":{},"f:nodeAffinity":{".":{},"f:requiredDuringSchedulingIgnoredDuringExecution":{}}},"f:containers":{"k:{\"name\":\"app\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:ports":{".":{},"k:{\"containerPort\":9376,\"protocol\":\"TCP\"}":{".":{},"f:containerPort":{},"f:protocol":{}}},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{},"f:tolerations":{}}}},{"manager":"kubelet","operation":"Update","apiVersion":"v1","time":"2023-02-25T11:28:39Z","fieldsType":"FieldsV1","fieldsV1":{"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.64.223\"}":{".":{},"f:ip":{}}},"f:startTime":{}}},"subresource":"status"}]},"spec":{"volumes":[{"name":"kube-api-access-p7qkn","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"app","image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","ports":[{"containerPort":9376,"protocol":"TCP"}],"resources":{},"volumeMounts":[{"name":"kube-api-access-p7qkn","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"File","imagePullPolicy":"IfNotPresent","securityContext":{}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","serviceAccountName":"default","serviceAccount":"default","nodeName":"diek3dua9gay-2","securityContext":{},"affinity":{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchFields":[{"key":"metadata.name","operator":"In","values":["diek3dua9gay-2"]}]}]}}},"schedulerName":"default-scheduler","tolerations":[{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/disk-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/pid-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unschedulable","operator":"Exists","effect":"NoSchedule"}],"priority":0,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:37Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:39Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:39Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:37Z"}],"hostIP":"192.168.121.117","podIP":"10.233.64.223","podIPs":[{"ip":"10.233.64.223"}],"startTime":"2023-02-25T11:28:37Z","containerStatuses":[{"name":"app","state":{"running":{"startedAt":"2023-02-25T11:28:38Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","imageID":"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3","containerID":"cri-o://6024590c52b6d7cc0372abb02e11c0c3eb487e06eac17c7559696d9c121d2c96","started":true}],"qosClass":"BestEffort"}},{"metadata":{"name":"daemon-set-k7fsc","generateName":"daemon-set-","namespace":"daemonsets-8281","uid":"5fff7f57-4a9f-427f-8aa7-9d747ac7487c","resourceVersion":"42805","creationTimestamp":"2023-02-25T11:28:37Z","deletionTimestamp":"2023-02-25T11:29:11Z","deletionGracePeriodSeconds":30,"labels":{"controller-revision-hash":"7f7ffb4fcc","daemonset-name":"daemon-set","pod-template-generation":"1"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"DaemonSet","name":"daemon-set","uid":"ba66bb7e-57bb-4b5c-9258-901aa735d014","controller":true,"blockOwnerDeletion":true}],"managedFields":[{"manager":"kube-controller-manager","operation":"Update","apiVersion":"v1","time":"2023-02-25T11:28:37Z","fieldsType":"FieldsV1","fieldsV1":{"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:controller-revision-hash":{},"f:daemonset-name":{},"f:pod-template-generation":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"ba66bb7e-57bb-4b5c-9258-901aa735d014\"}":{}}},"f:spec":{"f:affinity":{".":{},"f:nodeAffinity":{".":{},"f:requiredDuringSchedulingIgnoredDuringExecution":{}}},"f:containers":{"k:{\"name\":\"app\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:ports":{".":{},"k:{\"containerPort\":9376,\"protocol\":\"TCP\"}":{".":{},"f:containerPort":{},"f:protocol":{}}},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{},"f:tolerations":{}}}},{"manager":"kubelet","operation":"Update","apiVersion":"v1","time":"2023-02-25T11:28:40Z","fieldsType":"FieldsV1","fieldsV1":{"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.65.141\"}":{".":{},"f:ip":{}}},"f:startTime":{}}},"subresource":"status"}]},"spec":{"volumes":[{"name":"kube-api-access-j5mdp","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"app","image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","ports":[{"containerPort":9376,"protocol":"TCP"}],"resources":{},"volumeMounts":[{"name":"kube-api-access-j5mdp","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"File","imagePullPolicy":"IfNotPresent","securityContext":{}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","serviceAccountName":"default","serviceAccount":"default","nodeName":"diek3dua9gay-1","securityContext":{},"affinity":{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchFields":[{"key":"metadata.name","operator":"In","values":["diek3dua9gay-1"]}]}]}}},"schedulerName":"default-scheduler","tolerations":[{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/disk-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/pid-pressure","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unschedulable","operator":"Exists","effect":"NoSchedule"}],"priority":0,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:37Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:40Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:40Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2023-02-25T11:28:37Z"}],"hostIP":"192.168.121.250","podIP":"10.233.65.141","podIPs":[{"ip":"10.233.65.141"}],"startTime":"2023-02-25T11:28:37Z","containerStatuses":[{"name":"app","state":{"running":{"startedAt":"2023-02-25T11:28:39Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"registry.k8s.io/e2e-test-images/httpd:2.4.38-2","imageID":"registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3","containerID":"cri-o://20b693ee8e48492d54ffb578dd372904c7a2341a34c2f3d1a883e71bad1ce2be","started":true}],"qosClass":"BestEffort"}}]} + + [AfterEach] [sig-apps] Daemon set [Serial] test/e2e/framework/framework.go:187 - Dec 14 16:59:47.462: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "endpointslice-9032" for this suite. 12/14/22 16:59:47.47 + Feb 25 11:28:41.261: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "daemonsets-8281" for this suite. 02/25/23 11:28:41.273 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSS +SSSSSSSSSS ------------------------------ -[sig-node] RuntimeClass - should schedule a Pod requesting a RuntimeClass and initialize its Overhead [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:129 -[BeforeEach] [sig-node] RuntimeClass +[sig-cli] Kubectl client Kubectl api-versions + should check if v1 is in available api versions [Conformance] + test/e2e/kubectl/kubectl.go:822 +[BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:59:47.48 -Dec 14 16:59:47.480: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename runtimeclass 12/14/22 16:59:47.484 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:59:47.517 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:59:47.521 -[It] should schedule a Pod requesting a RuntimeClass and initialize its Overhead [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:129 -Dec 14 16:59:47.541: INFO: Waiting up to 1m20s for at least 1 pods in namespace runtimeclass-7638 to be scheduled -Dec 14 16:59:47.545: INFO: 1 pods are not scheduled: [runtimeclass-7638/test-runtimeclass-runtimeclass-7638-preconfigured-handler-8thbp(b525d556-9219-43ae-b534-2bdd3d4d1ce1)] -[AfterEach] [sig-node] RuntimeClass +STEP: Creating a kubernetes client 02/25/23 11:28:41.296 +Feb 25 11:28:41.296: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 11:28:41.3 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:41.396 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:41.401 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 +[It] should check if v1 is in available api versions [Conformance] + test/e2e/kubectl/kubectl.go:822 +STEP: validating api versions 02/25/23 11:28:41.405 +Feb 25 11:28:41.405: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-8739 api-versions' +Feb 25 11:28:41.538: INFO: stderr: "" +Feb 25 11:28:41.538: INFO: stdout: "admissionregistration.k8s.io/v1\napiextensions.k8s.io/v1\napiregistration.k8s.io/v1\napps/v1\nauthentication.k8s.io/v1\nauthorization.k8s.io/v1\nautoscaling/v1\nautoscaling/v2\nautoscaling/v2beta2\nbatch/v1\ncertificates.k8s.io/v1\ncilium.io/v2\ncilium.io/v2alpha1\ncoordination.k8s.io/v1\ndiscovery.k8s.io/v1\nevents.k8s.io/v1\nflowcontrol.apiserver.k8s.io/v1beta1\nflowcontrol.apiserver.k8s.io/v1beta2\nnetworking.k8s.io/v1\nnode.k8s.io/v1\npolicy/v1\nrbac.authorization.k8s.io/v1\nscheduling.k8s.io/v1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n" +[AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 -Dec 14 16:59:49.564: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "runtimeclass-7638" for this suite. 12/14/22 16:59:49.572 -{"msg":"PASSED [sig-node] RuntimeClass should schedule a Pod requesting a RuntimeClass and initialize its Overhead [NodeConformance] [Conformance]","completed":305,"skipped":5798,"failed":0} +Feb 25 11:28:41.538: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-8739" for this suite. 02/25/23 11:28:41.547 +{"msg":"PASSED [sig-cli] Kubectl client Kubectl api-versions should check if v1 is in available api versions [Conformance]","completed":301,"skipped":5776,"failed":0} ------------------------------ -• [2.100 seconds] -[sig-node] RuntimeClass -test/e2e/common/node/framework.go:23 - should schedule a Pod requesting a RuntimeClass and initialize its Overhead [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:129 +• [0.262 seconds] +[sig-cli] Kubectl client +test/e2e/kubectl/framework.go:23 + Kubectl api-versions + test/e2e/kubectl/kubectl.go:816 + should check if v1 is in available api versions [Conformance] + test/e2e/kubectl/kubectl.go:822 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] RuntimeClass + [BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:59:47.48 - Dec 14 16:59:47.480: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename runtimeclass 12/14/22 16:59:47.484 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:59:47.517 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:59:47.521 - [It] should schedule a Pod requesting a RuntimeClass and initialize its Overhead [NodeConformance] [Conformance] - test/e2e/common/node/runtimeclass.go:129 - Dec 14 16:59:47.541: INFO: Waiting up to 1m20s for at least 1 pods in namespace runtimeclass-7638 to be scheduled - Dec 14 16:59:47.545: INFO: 1 pods are not scheduled: [runtimeclass-7638/test-runtimeclass-runtimeclass-7638-preconfigured-handler-8thbp(b525d556-9219-43ae-b534-2bdd3d4d1ce1)] - [AfterEach] [sig-node] RuntimeClass + STEP: Creating a kubernetes client 02/25/23 11:28:41.296 + Feb 25 11:28:41.296: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 11:28:41.3 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:41.396 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:41.401 + [BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 + [It] should check if v1 is in available api versions [Conformance] + test/e2e/kubectl/kubectl.go:822 + STEP: validating api versions 02/25/23 11:28:41.405 + Feb 25 11:28:41.405: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-8739 api-versions' + Feb 25 11:28:41.538: INFO: stderr: "" + Feb 25 11:28:41.538: INFO: stdout: "admissionregistration.k8s.io/v1\napiextensions.k8s.io/v1\napiregistration.k8s.io/v1\napps/v1\nauthentication.k8s.io/v1\nauthorization.k8s.io/v1\nautoscaling/v1\nautoscaling/v2\nautoscaling/v2beta2\nbatch/v1\ncertificates.k8s.io/v1\ncilium.io/v2\ncilium.io/v2alpha1\ncoordination.k8s.io/v1\ndiscovery.k8s.io/v1\nevents.k8s.io/v1\nflowcontrol.apiserver.k8s.io/v1beta1\nflowcontrol.apiserver.k8s.io/v1beta2\nnetworking.k8s.io/v1\nnode.k8s.io/v1\npolicy/v1\nrbac.authorization.k8s.io/v1\nscheduling.k8s.io/v1\nstorage.k8s.io/v1\nstorage.k8s.io/v1beta1\nv1\n" + [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 16:59:49.564: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "runtimeclass-7638" for this suite. 12/14/22 16:59:49.572 + Feb 25 11:28:41.538: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-8739" for this suite. 02/25/23 11:28:41.547 << End Captured GinkgoWriter Output ------------------------------ -SS +SSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Ephemeral Containers [NodeConformance] - will start an ephemeral container in an existing pod [Conformance] - test/e2e/common/node/ephemeral_containers.go:45 -[BeforeEach] [sig-node] Ephemeral Containers [NodeConformance] +[sig-cli] Kubectl client Proxy server + should support --unix-socket=/path [Conformance] + test/e2e/kubectl/kubectl.go:1810 +[BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:59:49.582 -Dec 14 16:59:49.582: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename ephemeral-containers-test 12/14/22 16:59:49.585 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:59:49.605 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:59:49.609 -[BeforeEach] [sig-node] Ephemeral Containers [NodeConformance] - test/e2e/common/node/ephemeral_containers.go:38 -[It] will start an ephemeral container in an existing pod [Conformance] - test/e2e/common/node/ephemeral_containers.go:45 -STEP: creating a target pod 12/14/22 16:59:49.612 -Dec 14 16:59:49.625: INFO: Waiting up to 5m0s for pod "ephemeral-containers-target-pod" in namespace "ephemeral-containers-test-7151" to be "running and ready" -Dec 14 16:59:49.631: INFO: Pod "ephemeral-containers-target-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 5.990803ms -Dec 14 16:59:49.631: INFO: The phase of Pod ephemeral-containers-target-pod is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:59:51.640: INFO: Pod "ephemeral-containers-target-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015503693s -Dec 14 16:59:51.641: INFO: The phase of Pod ephemeral-containers-target-pod is Pending, waiting for it to be Running (with Ready = true) -Dec 14 16:59:53.639: INFO: Pod "ephemeral-containers-target-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.013877687s -Dec 14 16:59:53.639: INFO: The phase of Pod ephemeral-containers-target-pod is Running (Ready = true) -Dec 14 16:59:53.639: INFO: Pod "ephemeral-containers-target-pod" satisfied condition "running and ready" -STEP: adding an ephemeral container 12/14/22 16:59:53.644 -Dec 14 16:59:53.663: INFO: Waiting up to 1m0s for pod "ephemeral-containers-target-pod" in namespace "ephemeral-containers-test-7151" to be "container debugger running" -Dec 14 16:59:53.667: INFO: Pod "ephemeral-containers-target-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.505401ms -Dec 14 16:59:55.675: INFO: Pod "ephemeral-containers-target-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.012629606s -Dec 14 16:59:55.675: INFO: Pod "ephemeral-containers-target-pod" satisfied condition "container debugger running" -STEP: checking pod container endpoints 12/14/22 16:59:55.675 -Dec 14 16:59:55.677: INFO: ExecWithOptions {Command:[/bin/echo marco] Namespace:ephemeral-containers-test-7151 PodName:ephemeral-containers-target-pod ContainerName:debugger Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 16:59:55.677: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 16:59:55.680: INFO: ExecWithOptions: Clientset creation -Dec 14 16:59:55.680: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/ephemeral-containers-test-7151/pods/ephemeral-containers-target-pod/exec?command=%2Fbin%2Fecho&command=marco&container=debugger&container=debugger&stderr=true&stdout=true) -Dec 14 16:59:55.853: INFO: Exec stderr: "" -[AfterEach] [sig-node] Ephemeral Containers [NodeConformance] +STEP: Creating a kubernetes client 02/25/23 11:28:41.563 +Feb 25 11:28:41.563: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 11:28:41.565 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:41.602 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:41.606 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 +[It] should support --unix-socket=/path [Conformance] + test/e2e/kubectl/kubectl.go:1810 +STEP: Starting the proxy 02/25/23 11:28:41.611 +Feb 25 11:28:41.614: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6083 proxy --unix-socket=/tmp/kubectl-proxy-unix656268453/test' +STEP: retrieving proxy /api/ output 02/25/23 11:28:41.753 +[AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 -Dec 14 16:59:56.186: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "ephemeral-containers-test-7151" for this suite. 12/14/22 16:59:56.203 -{"msg":"PASSED [sig-node] Ephemeral Containers [NodeConformance] will start an ephemeral container in an existing pod [Conformance]","completed":306,"skipped":5800,"failed":0} +Feb 25 11:28:41.755: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-6083" for this suite. 02/25/23 11:28:41.766 +{"msg":"PASSED [sig-cli] Kubectl client Proxy server should support --unix-socket=/path [Conformance]","completed":302,"skipped":5803,"failed":0} ------------------------------ -• [SLOW TEST] [6.636 seconds] -[sig-node] Ephemeral Containers [NodeConformance] -test/e2e/common/node/framework.go:23 - will start an ephemeral container in an existing pod [Conformance] - test/e2e/common/node/ephemeral_containers.go:45 +• [0.237 seconds] +[sig-cli] Kubectl client +test/e2e/kubectl/framework.go:23 + Proxy server + test/e2e/kubectl/kubectl.go:1778 + should support --unix-socket=/path [Conformance] + test/e2e/kubectl/kubectl.go:1810 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Ephemeral Containers [NodeConformance] + [BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:59:49.582 - Dec 14 16:59:49.582: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename ephemeral-containers-test 12/14/22 16:59:49.585 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:59:49.605 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:59:49.609 - [BeforeEach] [sig-node] Ephemeral Containers [NodeConformance] - test/e2e/common/node/ephemeral_containers.go:38 - [It] will start an ephemeral container in an existing pod [Conformance] - test/e2e/common/node/ephemeral_containers.go:45 - STEP: creating a target pod 12/14/22 16:59:49.612 - Dec 14 16:59:49.625: INFO: Waiting up to 5m0s for pod "ephemeral-containers-target-pod" in namespace "ephemeral-containers-test-7151" to be "running and ready" - Dec 14 16:59:49.631: INFO: Pod "ephemeral-containers-target-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 5.990803ms - Dec 14 16:59:49.631: INFO: The phase of Pod ephemeral-containers-target-pod is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:59:51.640: INFO: Pod "ephemeral-containers-target-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015503693s - Dec 14 16:59:51.641: INFO: The phase of Pod ephemeral-containers-target-pod is Pending, waiting for it to be Running (with Ready = true) - Dec 14 16:59:53.639: INFO: Pod "ephemeral-containers-target-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.013877687s - Dec 14 16:59:53.639: INFO: The phase of Pod ephemeral-containers-target-pod is Running (Ready = true) - Dec 14 16:59:53.639: INFO: Pod "ephemeral-containers-target-pod" satisfied condition "running and ready" - STEP: adding an ephemeral container 12/14/22 16:59:53.644 - Dec 14 16:59:53.663: INFO: Waiting up to 1m0s for pod "ephemeral-containers-target-pod" in namespace "ephemeral-containers-test-7151" to be "container debugger running" - Dec 14 16:59:53.667: INFO: Pod "ephemeral-containers-target-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.505401ms - Dec 14 16:59:55.675: INFO: Pod "ephemeral-containers-target-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.012629606s - Dec 14 16:59:55.675: INFO: Pod "ephemeral-containers-target-pod" satisfied condition "container debugger running" - STEP: checking pod container endpoints 12/14/22 16:59:55.675 - Dec 14 16:59:55.677: INFO: ExecWithOptions {Command:[/bin/echo marco] Namespace:ephemeral-containers-test-7151 PodName:ephemeral-containers-target-pod ContainerName:debugger Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 16:59:55.677: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 16:59:55.680: INFO: ExecWithOptions: Clientset creation - Dec 14 16:59:55.680: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/ephemeral-containers-test-7151/pods/ephemeral-containers-target-pod/exec?command=%2Fbin%2Fecho&command=marco&container=debugger&container=debugger&stderr=true&stdout=true) - Dec 14 16:59:55.853: INFO: Exec stderr: "" - [AfterEach] [sig-node] Ephemeral Containers [NodeConformance] + STEP: Creating a kubernetes client 02/25/23 11:28:41.563 + Feb 25 11:28:41.563: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 11:28:41.565 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:41.602 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:41.606 + [BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 + [It] should support --unix-socket=/path [Conformance] + test/e2e/kubectl/kubectl.go:1810 + STEP: Starting the proxy 02/25/23 11:28:41.611 + Feb 25 11:28:41.614: INFO: Asynchronously running '/usr/local/bin/kubectl kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-6083 proxy --unix-socket=/tmp/kubectl-proxy-unix656268453/test' + STEP: retrieving proxy /api/ output 02/25/23 11:28:41.753 + [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 16:59:56.186: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "ephemeral-containers-test-7151" for this suite. 12/14/22 16:59:56.203 + Feb 25 11:28:41.755: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-6083" for this suite. 02/25/23 11:28:41.766 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Secrets - should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:88 -[BeforeEach] [sig-storage] Secrets +[sig-node] Container Runtime blackbox test on terminated container + should report termination message as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:231 +[BeforeEach] [sig-node] Container Runtime test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 16:59:56.224 -Dec 14 16:59:56.224: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename secrets 12/14/22 16:59:56.226 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:59:56.311 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:59:56.316 -[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:88 -STEP: Creating secret with name secret-test-map-3fdb6000-65ca-4228-83c9-27680d65be9b 12/14/22 16:59:56.32 -STEP: Creating a pod to test consume secrets 12/14/22 16:59:56.329 -Dec 14 16:59:56.344: INFO: Waiting up to 5m0s for pod "pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616" in namespace "secrets-8768" to be "Succeeded or Failed" -Dec 14 16:59:56.349: INFO: Pod "pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616": Phase="Pending", Reason="", readiness=false. Elapsed: 5.109007ms -Dec 14 16:59:58.356: INFO: Pod "pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011890822s -Dec 14 17:00:00.356: INFO: Pod "pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011410993s -STEP: Saw pod success 12/14/22 17:00:00.356 -Dec 14 17:00:00.357: INFO: Pod "pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616" satisfied condition "Succeeded or Failed" -Dec 14 17:00:00.361: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616 container secret-volume-test: -STEP: delete the pod 12/14/22 17:00:00.372 -Dec 14 17:00:00.391: INFO: Waiting for pod pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616 to disappear -Dec 14 17:00:00.397: INFO: Pod pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616 no longer exists -[AfterEach] [sig-storage] Secrets +STEP: Creating a kubernetes client 02/25/23 11:28:41.802 +Feb 25 11:28:41.802: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-runtime 02/25/23 11:28:41.81 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:41.851 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:41.857 +[It] should report termination message as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:231 +STEP: create the container 02/25/23 11:28:41.869 +STEP: wait for the container to reach Succeeded 02/25/23 11:28:41.883 +STEP: get the container status 02/25/23 11:28:49.978 +STEP: the container should be terminated 02/25/23 11:28:49.983 +STEP: the termination message should be set 02/25/23 11:28:49.983 +Feb 25 11:28:49.984: INFO: Expected: &{} to match Container's Termination Message: -- +STEP: delete the container 02/25/23 11:28:49.984 +[AfterEach] [sig-node] Container Runtime test/e2e/framework/framework.go:187 -Dec 14 17:00:00.397: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-8768" for this suite. 12/14/22 17:00:00.405 -{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","completed":307,"skipped":5838,"failed":0} +Feb 25 11:28:50.009: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-runtime-9603" for this suite. 02/25/23 11:28:50.015 +{"msg":"PASSED [sig-node] Container Runtime blackbox test on terminated container should report termination message as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","completed":303,"skipped":5831,"failed":0} ------------------------------ -• [4.194 seconds] -[sig-storage] Secrets -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:88 +• [SLOW TEST] [8.222 seconds] +[sig-node] Container Runtime +test/e2e/common/node/framework.go:23 + blackbox test + test/e2e/common/node/runtime.go:43 + on terminated container + test/e2e/common/node/runtime.go:136 + should report termination message as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:231 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Secrets + [BeforeEach] [sig-node] Container Runtime test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 16:59:56.224 - Dec 14 16:59:56.224: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename secrets 12/14/22 16:59:56.226 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 16:59:56.311 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 16:59:56.316 - [It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:88 - STEP: Creating secret with name secret-test-map-3fdb6000-65ca-4228-83c9-27680d65be9b 12/14/22 16:59:56.32 - STEP: Creating a pod to test consume secrets 12/14/22 16:59:56.329 - Dec 14 16:59:56.344: INFO: Waiting up to 5m0s for pod "pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616" in namespace "secrets-8768" to be "Succeeded or Failed" - Dec 14 16:59:56.349: INFO: Pod "pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616": Phase="Pending", Reason="", readiness=false. Elapsed: 5.109007ms - Dec 14 16:59:58.356: INFO: Pod "pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011890822s - Dec 14 17:00:00.356: INFO: Pod "pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.011410993s - STEP: Saw pod success 12/14/22 17:00:00.356 - Dec 14 17:00:00.357: INFO: Pod "pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616" satisfied condition "Succeeded or Failed" - Dec 14 17:00:00.361: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616 container secret-volume-test: - STEP: delete the pod 12/14/22 17:00:00.372 - Dec 14 17:00:00.391: INFO: Waiting for pod pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616 to disappear - Dec 14 17:00:00.397: INFO: Pod pod-secrets-e29ba9f9-30b0-4380-ab71-a7bb43e8f616 no longer exists - [AfterEach] [sig-storage] Secrets + STEP: Creating a kubernetes client 02/25/23 11:28:41.802 + Feb 25 11:28:41.802: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-runtime 02/25/23 11:28:41.81 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:41.851 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:41.857 + [It] should report termination message as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:231 + STEP: create the container 02/25/23 11:28:41.869 + STEP: wait for the container to reach Succeeded 02/25/23 11:28:41.883 + STEP: get the container status 02/25/23 11:28:49.978 + STEP: the container should be terminated 02/25/23 11:28:49.983 + STEP: the termination message should be set 02/25/23 11:28:49.983 + Feb 25 11:28:49.984: INFO: Expected: &{} to match Container's Termination Message: -- + STEP: delete the container 02/25/23 11:28:49.984 + [AfterEach] [sig-node] Container Runtime test/e2e/framework/framework.go:187 - Dec 14 17:00:00.397: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "secrets-8768" for this suite. 12/14/22 17:00:00.405 + Feb 25 11:28:50.009: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-runtime-9603" for this suite. 02/25/23 11:28:50.015 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] - works for multiple CRDs of different groups [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:275 + works for CRD preserving unknown fields in an embedded object [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:235 [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:00:00.42 -Dec 14 17:00:00.420: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 17:00:00.427 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:00:00.463 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:00:00.468 -[It] works for multiple CRDs of different groups [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:275 -STEP: CRs in different groups (two CRDs) show up in OpenAPI documentation 12/14/22 17:00:00.472 -Dec 14 17:00:00.473: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 17:00:03.958: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 +STEP: Creating a kubernetes client 02/25/23 11:28:50.037 +Feb 25 11:28:50.038: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 11:28:50.039 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:50.068 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:50.071 +[It] works for CRD preserving unknown fields in an embedded object [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:235 +Feb 25 11:28:50.076: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: kubectl validation (kubectl create and apply) allows request with any unknown properties 02/25/23 11:28:56.446 +Feb 25 11:28:56.447: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-8721 --namespace=crd-publish-openapi-8721 create -f -' +Feb 25 11:28:58.219: INFO: stderr: "" +Feb 25 11:28:58.219: INFO: stdout: "e2e-test-crd-publish-openapi-1882-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n" +Feb 25 11:28:58.223: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-8721 --namespace=crd-publish-openapi-8721 delete e2e-test-crd-publish-openapi-1882-crds test-cr' +Feb 25 11:28:58.420: INFO: stderr: "" +Feb 25 11:28:58.420: INFO: stdout: "e2e-test-crd-publish-openapi-1882-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n" +Feb 25 11:28:58.421: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-8721 --namespace=crd-publish-openapi-8721 apply -f -' +Feb 25 11:28:59.090: INFO: stderr: "" +Feb 25 11:28:59.090: INFO: stdout: "e2e-test-crd-publish-openapi-1882-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n" +Feb 25 11:28:59.090: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-8721 --namespace=crd-publish-openapi-8721 delete e2e-test-crd-publish-openapi-1882-crds test-cr' +Feb 25 11:28:59.310: INFO: stderr: "" +Feb 25 11:28:59.310: INFO: stdout: "e2e-test-crd-publish-openapi-1882-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n" +STEP: kubectl explain works to explain CR 02/25/23 11:28:59.31 +Feb 25 11:28:59.311: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-8721 explain e2e-test-crd-publish-openapi-1882-crds' +Feb 25 11:28:59.791: INFO: stderr: "" +Feb 25 11:28:59.791: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-1882-crd\nVERSION: crd-publish-openapi-test-unknown-in-nested.example.com/v1\n\nDESCRIPTION:\n preserve-unknown-properties in nested field for Testing\n\nFIELDS:\n apiVersion\t\n APIVersion defines the versioned schema of this representation of an\n object. Servers should convert recognized schemas to the latest internal\n value, and may reject unrecognized values. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n kind\t\n Kind is a string value representing the REST resource this object\n represents. Servers may infer this from the endpoint the client submits\n requests to. Cannot be updated. In CamelCase. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n metadata\t\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n spec\t<>\n Specification of Waldo\n\n status\t\n Status of Waldo\n\n" [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 17:00:17.815: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "crd-publish-openapi-4862" for this suite. 12/14/22 17:00:17.828 -{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of different groups [Conformance]","completed":308,"skipped":5851,"failed":0} +Feb 25 11:29:03.777: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "crd-publish-openapi-8721" for this suite. 02/25/23 11:29:03.792 +{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for CRD preserving unknown fields in an embedded object [Conformance]","completed":304,"skipped":5859,"failed":0} ------------------------------ -• [SLOW TEST] [17.417 seconds] +• [SLOW TEST] [13.765 seconds] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/apimachinery/framework.go:23 - works for multiple CRDs of different groups [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:275 + works for CRD preserving unknown fields in an embedded object [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:235 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:00:00.42 - Dec 14 17:00:00.420: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename crd-publish-openapi 12/14/22 17:00:00.427 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:00:00.463 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:00:00.468 - [It] works for multiple CRDs of different groups [Conformance] - test/e2e/apimachinery/crd_publish_openapi.go:275 - STEP: CRs in different groups (two CRDs) show up in OpenAPI documentation 12/14/22 17:00:00.472 - Dec 14 17:00:00.473: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 17:00:03.958: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 + STEP: Creating a kubernetes client 02/25/23 11:28:50.037 + Feb 25 11:28:50.038: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 11:28:50.039 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:28:50.068 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:28:50.071 + [It] works for CRD preserving unknown fields in an embedded object [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:235 + Feb 25 11:28:50.076: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: kubectl validation (kubectl create and apply) allows request with any unknown properties 02/25/23 11:28:56.446 + Feb 25 11:28:56.447: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-8721 --namespace=crd-publish-openapi-8721 create -f -' + Feb 25 11:28:58.219: INFO: stderr: "" + Feb 25 11:28:58.219: INFO: stdout: "e2e-test-crd-publish-openapi-1882-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n" + Feb 25 11:28:58.223: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-8721 --namespace=crd-publish-openapi-8721 delete e2e-test-crd-publish-openapi-1882-crds test-cr' + Feb 25 11:28:58.420: INFO: stderr: "" + Feb 25 11:28:58.420: INFO: stdout: "e2e-test-crd-publish-openapi-1882-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n" + Feb 25 11:28:58.421: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-8721 --namespace=crd-publish-openapi-8721 apply -f -' + Feb 25 11:28:59.090: INFO: stderr: "" + Feb 25 11:28:59.090: INFO: stdout: "e2e-test-crd-publish-openapi-1882-crd.crd-publish-openapi-test-unknown-in-nested.example.com/test-cr created\n" + Feb 25 11:28:59.090: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-8721 --namespace=crd-publish-openapi-8721 delete e2e-test-crd-publish-openapi-1882-crds test-cr' + Feb 25 11:28:59.310: INFO: stderr: "" + Feb 25 11:28:59.310: INFO: stdout: "e2e-test-crd-publish-openapi-1882-crd.crd-publish-openapi-test-unknown-in-nested.example.com \"test-cr\" deleted\n" + STEP: kubectl explain works to explain CR 02/25/23 11:28:59.31 + Feb 25 11:28:59.311: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=crd-publish-openapi-8721 explain e2e-test-crd-publish-openapi-1882-crds' + Feb 25 11:28:59.791: INFO: stderr: "" + Feb 25 11:28:59.791: INFO: stdout: "KIND: e2e-test-crd-publish-openapi-1882-crd\nVERSION: crd-publish-openapi-test-unknown-in-nested.example.com/v1\n\nDESCRIPTION:\n preserve-unknown-properties in nested field for Testing\n\nFIELDS:\n apiVersion\t\n APIVersion defines the versioned schema of this representation of an\n object. Servers should convert recognized schemas to the latest internal\n value, and may reject unrecognized values. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources\n\n kind\t\n Kind is a string value representing the REST resource this object\n represents. Servers may infer this from the endpoint the client submits\n requests to. Cannot be updated. In CamelCase. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds\n\n metadata\t\n Standard object's metadata. More info:\n https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\n\n spec\t<>\n Specification of Waldo\n\n status\t\n Status of Waldo\n\n" [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 17:00:17.815: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "crd-publish-openapi-4862" for this suite. 12/14/22 17:00:17.828 + Feb 25 11:29:03.777: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "crd-publish-openapi-8721" for this suite. 02/25/23 11:29:03.792 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSS +SSSSSSSSSSSSSSSS ------------------------------ [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - patching/updating a validating webhook should work [Conformance] - test/e2e/apimachinery/webhook.go:412 + should be able to deny pod and configmap creation [Conformance] + test/e2e/apimachinery/webhook.go:196 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:00:17.84 -Dec 14 17:00:17.840: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 17:00:17.842 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:00:17.9 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:00:17.903 +STEP: Creating a kubernetes client 02/25/23 11:29:03.804 +Feb 25 11:29:03.804: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 11:29:03.807 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:29:03.833 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:29:03.837 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 17:00:17.923 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 17:00:18.869 -STEP: Deploying the webhook pod 12/14/22 17:00:18.906 -STEP: Wait for the deployment to be ready 12/14/22 17:00:18.922 -Dec 14 17:00:18.931: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created -STEP: Deploying the webhook service 12/14/22 17:00:20.95 -STEP: Verifying the service has paired with the endpoint 12/14/22 17:00:20.964 -Dec 14 17:00:21.965: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] patching/updating a validating webhook should work [Conformance] - test/e2e/apimachinery/webhook.go:412 -STEP: Creating a validating webhook configuration 12/14/22 17:00:21.97 -STEP: Creating a configMap that does not comply to the validation webhook rules 12/14/22 17:00:22.01 -STEP: Updating a validating webhook configuration's rules to not include the create operation 12/14/22 17:00:22.052 -STEP: Creating a configMap that does not comply to the validation webhook rules 12/14/22 17:00:22.077 -STEP: Patching a validating webhook configuration's rules to include the create operation 12/14/22 17:00:22.198 -STEP: Creating a configMap that does not comply to the validation webhook rules 12/14/22 17:00:22.206 +STEP: Setting up server cert 02/25/23 11:29:03.859 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 11:29:05.615 +STEP: Deploying the webhook pod 02/25/23 11:29:05.642 +STEP: Wait for the deployment to be ready 02/25/23 11:29:05.66 +Feb 25 11:29:05.685: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set +Feb 25 11:29:07.706: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 11, 29, 5, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 29, 5, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 11, 29, 5, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 29, 5, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} +STEP: Deploying the webhook service 02/25/23 11:29:09.713 +STEP: Verifying the service has paired with the endpoint 02/25/23 11:29:09.737 +Feb 25 11:29:10.737: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should be able to deny pod and configmap creation [Conformance] + test/e2e/apimachinery/webhook.go:196 +STEP: Registering the webhook via the AdmissionRegistration API 02/25/23 11:29:10.744 +STEP: create a pod that should be denied by the webhook 02/25/23 11:29:10.775 +STEP: create a pod that causes the webhook to hang 02/25/23 11:29:10.799 +STEP: create a configmap that should be denied by the webhook 02/25/23 11:29:20.81 +STEP: create a configmap that should be admitted by the webhook 02/25/23 11:29:20.855 +STEP: update (PUT) the admitted configmap to a non-compliant one should be rejected by the webhook 02/25/23 11:29:20.901 +STEP: update (PATCH) the admitted configmap to a non-compliant one should be rejected by the webhook 02/25/23 11:29:20.931 +STEP: create a namespace that bypass the webhook 02/25/23 11:29:20.95 +STEP: create a configmap that violates the webhook policy but is in a whitelisted namespace 02/25/23 11:29:20.973 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 17:00:22.216: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-1376" for this suite. 12/14/22 17:00:22.221 -STEP: Destroying namespace "webhook-1376-markers" for this suite. 12/14/22 17:00:22.229 +Feb 25 11:29:21.014: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-3580" for this suite. 02/25/23 11:29:21.025 +STEP: Destroying namespace "webhook-3580-markers" for this suite. 02/25/23 11:29:21.063 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] patching/updating a validating webhook should work [Conformance]","completed":309,"skipped":5862,"failed":0} +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny pod and configmap creation [Conformance]","completed":305,"skipped":5875,"failed":0} ------------------------------ -• [4.460 seconds] +• [SLOW TEST] [17.489 seconds] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/framework.go:23 - patching/updating a validating webhook should work [Conformance] - test/e2e/apimachinery/webhook.go:412 + should be able to deny pod and configmap creation [Conformance] + test/e2e/apimachinery/webhook.go:196 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:00:17.84 - Dec 14 17:00:17.840: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 17:00:17.842 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:00:17.9 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:00:17.903 + STEP: Creating a kubernetes client 02/25/23 11:29:03.804 + Feb 25 11:29:03.804: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 11:29:03.807 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:29:03.833 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:29:03.837 [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 17:00:17.923 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 17:00:18.869 - STEP: Deploying the webhook pod 12/14/22 17:00:18.906 - STEP: Wait for the deployment to be ready 12/14/22 17:00:18.922 - Dec 14 17:00:18.931: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created - STEP: Deploying the webhook service 12/14/22 17:00:20.95 - STEP: Verifying the service has paired with the endpoint 12/14/22 17:00:20.964 - Dec 14 17:00:21.965: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] patching/updating a validating webhook should work [Conformance] - test/e2e/apimachinery/webhook.go:412 - STEP: Creating a validating webhook configuration 12/14/22 17:00:21.97 - STEP: Creating a configMap that does not comply to the validation webhook rules 12/14/22 17:00:22.01 - STEP: Updating a validating webhook configuration's rules to not include the create operation 12/14/22 17:00:22.052 - STEP: Creating a configMap that does not comply to the validation webhook rules 12/14/22 17:00:22.077 - STEP: Patching a validating webhook configuration's rules to include the create operation 12/14/22 17:00:22.198 - STEP: Creating a configMap that does not comply to the validation webhook rules 12/14/22 17:00:22.206 + STEP: Setting up server cert 02/25/23 11:29:03.859 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 11:29:05.615 + STEP: Deploying the webhook pod 02/25/23 11:29:05.642 + STEP: Wait for the deployment to be ready 02/25/23 11:29:05.66 + Feb 25 11:29:05.685: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set + Feb 25 11:29:07.706: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 11, 29, 5, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 29, 5, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 11, 29, 5, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 29, 5, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} + STEP: Deploying the webhook service 02/25/23 11:29:09.713 + STEP: Verifying the service has paired with the endpoint 02/25/23 11:29:09.737 + Feb 25 11:29:10.737: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] should be able to deny pod and configmap creation [Conformance] + test/e2e/apimachinery/webhook.go:196 + STEP: Registering the webhook via the AdmissionRegistration API 02/25/23 11:29:10.744 + STEP: create a pod that should be denied by the webhook 02/25/23 11:29:10.775 + STEP: create a pod that causes the webhook to hang 02/25/23 11:29:10.799 + STEP: create a configmap that should be denied by the webhook 02/25/23 11:29:20.81 + STEP: create a configmap that should be admitted by the webhook 02/25/23 11:29:20.855 + STEP: update (PUT) the admitted configmap to a non-compliant one should be rejected by the webhook 02/25/23 11:29:20.901 + STEP: update (PATCH) the admitted configmap to a non-compliant one should be rejected by the webhook 02/25/23 11:29:20.931 + STEP: create a namespace that bypass the webhook 02/25/23 11:29:20.95 + STEP: create a configmap that violates the webhook policy but is in a whitelisted namespace 02/25/23 11:29:20.973 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 17:00:22.216: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-1376" for this suite. 12/14/22 17:00:22.221 - STEP: Destroying namespace "webhook-1376-markers" for this suite. 12/14/22 17:00:22.229 + Feb 25 11:29:21.014: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-3580" for this suite. 02/25/23 11:29:21.025 + STEP: Destroying namespace "webhook-3580-markers" for this suite. 02/25/23 11:29:21.063 [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/apimachinery/webhook.go:104 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook - should execute poststart http hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:130 -[BeforeEach] [sig-node] Container Lifecycle Hook +[sig-apps] DisruptionController + should update/patch PodDisruptionBudget status [Conformance] + test/e2e/apps/disruption.go:163 +[BeforeEach] [sig-apps] DisruptionController test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:00:22.302 -Dec 14 17:00:22.303: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-lifecycle-hook 12/14/22 17:00:22.304 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:00:22.361 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:00:22.365 -[BeforeEach] when create a pod with lifecycle hook - test/e2e/common/node/lifecycle_hook.go:55 -STEP: create the container to handle the HTTPGet hook request. 12/14/22 17:00:22.377 -Dec 14 17:00:22.402: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-3492" to be "running and ready" -Dec 14 17:00:22.410: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 7.341342ms -Dec 14 17:00:22.410: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) -Dec 14 17:00:24.419: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 2.017021954s -Dec 14 17:00:24.420: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) -Dec 14 17:00:24.420: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" -[It] should execute poststart http hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:130 -STEP: create the pod with lifecycle hook 12/14/22 17:00:24.425 -Dec 14 17:00:24.432: INFO: Waiting up to 5m0s for pod "pod-with-poststart-http-hook" in namespace "container-lifecycle-hook-3492" to be "running and ready" -Dec 14 17:00:24.439: INFO: Pod "pod-with-poststart-http-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 6.671694ms -Dec 14 17:00:24.439: INFO: The phase of Pod pod-with-poststart-http-hook is Pending, waiting for it to be Running (with Ready = true) -Dec 14 17:00:26.448: INFO: Pod "pod-with-poststart-http-hook": Phase="Running", Reason="", readiness=true. Elapsed: 2.016029034s -Dec 14 17:00:26.449: INFO: The phase of Pod pod-with-poststart-http-hook is Running (Ready = true) -Dec 14 17:00:26.449: INFO: Pod "pod-with-poststart-http-hook" satisfied condition "running and ready" -STEP: check poststart hook 12/14/22 17:00:26.453 -STEP: delete the pod with lifecycle hook 12/14/22 17:00:26.479 -Dec 14 17:00:26.487: INFO: Waiting for pod pod-with-poststart-http-hook to disappear -Dec 14 17:00:26.492: INFO: Pod pod-with-poststart-http-hook still exists -Dec 14 17:00:28.493: INFO: Waiting for pod pod-with-poststart-http-hook to disappear -Dec 14 17:00:28.500: INFO: Pod pod-with-poststart-http-hook still exists -Dec 14 17:00:30.493: INFO: Waiting for pod pod-with-poststart-http-hook to disappear -Dec 14 17:00:30.502: INFO: Pod pod-with-poststart-http-hook no longer exists -[AfterEach] [sig-node] Container Lifecycle Hook +STEP: Creating a kubernetes client 02/25/23 11:29:21.301 +Feb 25 11:29:21.301: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename disruption 02/25/23 11:29:21.304 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:29:21.416 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:29:21.421 +[BeforeEach] [sig-apps] DisruptionController + test/e2e/apps/disruption.go:71 +[It] should update/patch PodDisruptionBudget status [Conformance] + test/e2e/apps/disruption.go:163 +STEP: Waiting for the pdb to be processed 02/25/23 11:29:21.449 +STEP: Updating PodDisruptionBudget status 02/25/23 11:29:21.478 +STEP: Waiting for all pods to be running 02/25/23 11:29:21.542 +Feb 25 11:29:21.560: INFO: running pods: 0 < 1 +STEP: locating a running pod 02/25/23 11:29:23.566 +STEP: Waiting for the pdb to be processed 02/25/23 11:29:23.589 +STEP: Patching PodDisruptionBudget status 02/25/23 11:29:23.599 +STEP: Waiting for the pdb to be processed 02/25/23 11:29:23.636 +[AfterEach] [sig-apps] DisruptionController test/e2e/framework/framework.go:187 -Dec 14 17:00:30.502: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-lifecycle-hook-3492" for this suite. 12/14/22 17:00:30.51 -{"msg":"PASSED [sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute poststart http hook properly [NodeConformance] [Conformance]","completed":310,"skipped":5870,"failed":0} +Feb 25 11:29:23.643: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "disruption-7893" for this suite. 02/25/23 11:29:23.652 +{"msg":"PASSED [sig-apps] DisruptionController should update/patch PodDisruptionBudget status [Conformance]","completed":306,"skipped":5926,"failed":0} ------------------------------ -• [SLOW TEST] [8.218 seconds] -[sig-node] Container Lifecycle Hook -test/e2e/common/node/framework.go:23 - when create a pod with lifecycle hook - test/e2e/common/node/lifecycle_hook.go:46 - should execute poststart http hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:130 +• [2.361 seconds] +[sig-apps] DisruptionController +test/e2e/apps/framework.go:23 + should update/patch PodDisruptionBudget status [Conformance] + test/e2e/apps/disruption.go:163 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Container Lifecycle Hook + [BeforeEach] [sig-apps] DisruptionController test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:00:22.302 - Dec 14 17:00:22.303: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-lifecycle-hook 12/14/22 17:00:22.304 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:00:22.361 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:00:22.365 - [BeforeEach] when create a pod with lifecycle hook - test/e2e/common/node/lifecycle_hook.go:55 - STEP: create the container to handle the HTTPGet hook request. 12/14/22 17:00:22.377 - Dec 14 17:00:22.402: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-3492" to be "running and ready" - Dec 14 17:00:22.410: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 7.341342ms - Dec 14 17:00:22.410: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) - Dec 14 17:00:24.419: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 2.017021954s - Dec 14 17:00:24.420: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) - Dec 14 17:00:24.420: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" - [It] should execute poststart http hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:130 - STEP: create the pod with lifecycle hook 12/14/22 17:00:24.425 - Dec 14 17:00:24.432: INFO: Waiting up to 5m0s for pod "pod-with-poststart-http-hook" in namespace "container-lifecycle-hook-3492" to be "running and ready" - Dec 14 17:00:24.439: INFO: Pod "pod-with-poststart-http-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 6.671694ms - Dec 14 17:00:24.439: INFO: The phase of Pod pod-with-poststart-http-hook is Pending, waiting for it to be Running (with Ready = true) - Dec 14 17:00:26.448: INFO: Pod "pod-with-poststart-http-hook": Phase="Running", Reason="", readiness=true. Elapsed: 2.016029034s - Dec 14 17:00:26.449: INFO: The phase of Pod pod-with-poststart-http-hook is Running (Ready = true) - Dec 14 17:00:26.449: INFO: Pod "pod-with-poststart-http-hook" satisfied condition "running and ready" - STEP: check poststart hook 12/14/22 17:00:26.453 - STEP: delete the pod with lifecycle hook 12/14/22 17:00:26.479 - Dec 14 17:00:26.487: INFO: Waiting for pod pod-with-poststart-http-hook to disappear - Dec 14 17:00:26.492: INFO: Pod pod-with-poststart-http-hook still exists - Dec 14 17:00:28.493: INFO: Waiting for pod pod-with-poststart-http-hook to disappear - Dec 14 17:00:28.500: INFO: Pod pod-with-poststart-http-hook still exists - Dec 14 17:00:30.493: INFO: Waiting for pod pod-with-poststart-http-hook to disappear - Dec 14 17:00:30.502: INFO: Pod pod-with-poststart-http-hook no longer exists - [AfterEach] [sig-node] Container Lifecycle Hook + STEP: Creating a kubernetes client 02/25/23 11:29:21.301 + Feb 25 11:29:21.301: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename disruption 02/25/23 11:29:21.304 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:29:21.416 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:29:21.421 + [BeforeEach] [sig-apps] DisruptionController + test/e2e/apps/disruption.go:71 + [It] should update/patch PodDisruptionBudget status [Conformance] + test/e2e/apps/disruption.go:163 + STEP: Waiting for the pdb to be processed 02/25/23 11:29:21.449 + STEP: Updating PodDisruptionBudget status 02/25/23 11:29:21.478 + STEP: Waiting for all pods to be running 02/25/23 11:29:21.542 + Feb 25 11:29:21.560: INFO: running pods: 0 < 1 + STEP: locating a running pod 02/25/23 11:29:23.566 + STEP: Waiting for the pdb to be processed 02/25/23 11:29:23.589 + STEP: Patching PodDisruptionBudget status 02/25/23 11:29:23.599 + STEP: Waiting for the pdb to be processed 02/25/23 11:29:23.636 + [AfterEach] [sig-apps] DisruptionController test/e2e/framework/framework.go:187 - Dec 14 17:00:30.502: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-lifecycle-hook-3492" for this suite. 12/14/22 17:00:30.51 + Feb 25 11:29:23.643: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "disruption-7893" for this suite. 02/25/23 11:29:23.652 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-node] Container Lifecycle Hook when create a pod with lifecycle hook - should execute prestop http hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:152 -[BeforeEach] [sig-node] Container Lifecycle Hook +[sig-node] PodTemplates + should replace a pod template [Conformance] + test/e2e/common/node/podtemplates.go:176 +[BeforeEach] [sig-node] PodTemplates test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:00:30.526 -Dec 14 17:00:30.526: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-lifecycle-hook 12/14/22 17:00:30.528 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:00:30.563 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:00:30.57 -[BeforeEach] when create a pod with lifecycle hook - test/e2e/common/node/lifecycle_hook.go:55 -STEP: create the container to handle the HTTPGet hook request. 12/14/22 17:00:30.581 -Dec 14 17:00:30.596: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-4184" to be "running and ready" -Dec 14 17:00:30.601: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 5.39748ms -Dec 14 17:00:30.601: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) -Dec 14 17:00:32.610: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 2.013876206s -Dec 14 17:00:32.610: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) -Dec 14 17:00:32.610: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" -[It] should execute prestop http hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:152 -STEP: create the pod with lifecycle hook 12/14/22 17:00:32.614 -Dec 14 17:00:32.623: INFO: Waiting up to 5m0s for pod "pod-with-prestop-http-hook" in namespace "container-lifecycle-hook-4184" to be "running and ready" -Dec 14 17:00:32.627: INFO: Pod "pod-with-prestop-http-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 4.118022ms -Dec 14 17:00:32.628: INFO: The phase of Pod pod-with-prestop-http-hook is Pending, waiting for it to be Running (with Ready = true) -Dec 14 17:00:34.635: INFO: Pod "pod-with-prestop-http-hook": Phase="Running", Reason="", readiness=true. Elapsed: 2.012140149s -Dec 14 17:00:34.636: INFO: The phase of Pod pod-with-prestop-http-hook is Running (Ready = true) -Dec 14 17:00:34.636: INFO: Pod "pod-with-prestop-http-hook" satisfied condition "running and ready" -STEP: delete the pod with lifecycle hook 12/14/22 17:00:34.641 -Dec 14 17:00:34.654: INFO: Waiting for pod pod-with-prestop-http-hook to disappear -Dec 14 17:00:34.659: INFO: Pod pod-with-prestop-http-hook still exists -Dec 14 17:00:36.660: INFO: Waiting for pod pod-with-prestop-http-hook to disappear -Dec 14 17:00:36.671: INFO: Pod pod-with-prestop-http-hook still exists -Dec 14 17:00:38.660: INFO: Waiting for pod pod-with-prestop-http-hook to disappear -Dec 14 17:00:38.665: INFO: Pod pod-with-prestop-http-hook no longer exists -STEP: check prestop hook 12/14/22 17:00:38.665 -[AfterEach] [sig-node] Container Lifecycle Hook +STEP: Creating a kubernetes client 02/25/23 11:29:23.666 +Feb 25 11:29:23.666: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename podtemplate 02/25/23 11:29:23.668 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:29:23.717 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:29:23.722 +[It] should replace a pod template [Conformance] + test/e2e/common/node/podtemplates.go:176 +STEP: Create a pod template 02/25/23 11:29:23.728 +STEP: Replace a pod template 02/25/23 11:29:23.738 +Feb 25 11:29:23.759: INFO: Found updated podtemplate annotation: "true" + +[AfterEach] [sig-node] PodTemplates test/e2e/framework/framework.go:187 -Dec 14 17:00:38.681: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-lifecycle-hook-4184" for this suite. 12/14/22 17:00:38.689 -{"msg":"PASSED [sig-node] Container Lifecycle Hook when create a pod with lifecycle hook should execute prestop http hook properly [NodeConformance] [Conformance]","completed":311,"skipped":5901,"failed":0} +Feb 25 11:29:23.759: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "podtemplate-6243" for this suite. 02/25/23 11:29:23.767 +{"msg":"PASSED [sig-node] PodTemplates should replace a pod template [Conformance]","completed":307,"skipped":5961,"failed":0} ------------------------------ -• [SLOW TEST] [8.172 seconds] -[sig-node] Container Lifecycle Hook +• [0.112 seconds] +[sig-node] PodTemplates test/e2e/common/node/framework.go:23 - when create a pod with lifecycle hook - test/e2e/common/node/lifecycle_hook.go:46 - should execute prestop http hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:152 + should replace a pod template [Conformance] + test/e2e/common/node/podtemplates.go:176 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Container Lifecycle Hook + [BeforeEach] [sig-node] PodTemplates test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:00:30.526 - Dec 14 17:00:30.526: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-lifecycle-hook 12/14/22 17:00:30.528 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:00:30.563 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:00:30.57 - [BeforeEach] when create a pod with lifecycle hook - test/e2e/common/node/lifecycle_hook.go:55 - STEP: create the container to handle the HTTPGet hook request. 12/14/22 17:00:30.581 - Dec 14 17:00:30.596: INFO: Waiting up to 5m0s for pod "pod-handle-http-request" in namespace "container-lifecycle-hook-4184" to be "running and ready" - Dec 14 17:00:30.601: INFO: Pod "pod-handle-http-request": Phase="Pending", Reason="", readiness=false. Elapsed: 5.39748ms - Dec 14 17:00:30.601: INFO: The phase of Pod pod-handle-http-request is Pending, waiting for it to be Running (with Ready = true) - Dec 14 17:00:32.610: INFO: Pod "pod-handle-http-request": Phase="Running", Reason="", readiness=true. Elapsed: 2.013876206s - Dec 14 17:00:32.610: INFO: The phase of Pod pod-handle-http-request is Running (Ready = true) - Dec 14 17:00:32.610: INFO: Pod "pod-handle-http-request" satisfied condition "running and ready" - [It] should execute prestop http hook properly [NodeConformance] [Conformance] - test/e2e/common/node/lifecycle_hook.go:152 - STEP: create the pod with lifecycle hook 12/14/22 17:00:32.614 - Dec 14 17:00:32.623: INFO: Waiting up to 5m0s for pod "pod-with-prestop-http-hook" in namespace "container-lifecycle-hook-4184" to be "running and ready" - Dec 14 17:00:32.627: INFO: Pod "pod-with-prestop-http-hook": Phase="Pending", Reason="", readiness=false. Elapsed: 4.118022ms - Dec 14 17:00:32.628: INFO: The phase of Pod pod-with-prestop-http-hook is Pending, waiting for it to be Running (with Ready = true) - Dec 14 17:00:34.635: INFO: Pod "pod-with-prestop-http-hook": Phase="Running", Reason="", readiness=true. Elapsed: 2.012140149s - Dec 14 17:00:34.636: INFO: The phase of Pod pod-with-prestop-http-hook is Running (Ready = true) - Dec 14 17:00:34.636: INFO: Pod "pod-with-prestop-http-hook" satisfied condition "running and ready" - STEP: delete the pod with lifecycle hook 12/14/22 17:00:34.641 - Dec 14 17:00:34.654: INFO: Waiting for pod pod-with-prestop-http-hook to disappear - Dec 14 17:00:34.659: INFO: Pod pod-with-prestop-http-hook still exists - Dec 14 17:00:36.660: INFO: Waiting for pod pod-with-prestop-http-hook to disappear - Dec 14 17:00:36.671: INFO: Pod pod-with-prestop-http-hook still exists - Dec 14 17:00:38.660: INFO: Waiting for pod pod-with-prestop-http-hook to disappear - Dec 14 17:00:38.665: INFO: Pod pod-with-prestop-http-hook no longer exists - STEP: check prestop hook 12/14/22 17:00:38.665 - [AfterEach] [sig-node] Container Lifecycle Hook + STEP: Creating a kubernetes client 02/25/23 11:29:23.666 + Feb 25 11:29:23.666: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename podtemplate 02/25/23 11:29:23.668 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:29:23.717 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:29:23.722 + [It] should replace a pod template [Conformance] + test/e2e/common/node/podtemplates.go:176 + STEP: Create a pod template 02/25/23 11:29:23.728 + STEP: Replace a pod template 02/25/23 11:29:23.738 + Feb 25 11:29:23.759: INFO: Found updated podtemplate annotation: "true" + + [AfterEach] [sig-node] PodTemplates + test/e2e/framework/framework.go:187 + Feb 25 11:29:23.759: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "podtemplate-6243" for this suite. 02/25/23 11:29:23.767 + << End Captured GinkgoWriter Output +------------------------------ +[sig-node] Pods + should contain environment variables for services [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:443 +[BeforeEach] [sig-node] Pods + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 11:29:23.778 +Feb 25 11:29:23.778: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename pods 02/25/23 11:29:23.78 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:29:23.809 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:29:23.813 +[BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 +[It] should contain environment variables for services [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:443 +Feb 25 11:29:23.837: INFO: Waiting up to 5m0s for pod "server-envvars-cbbf4662-3990-4879-8518-62c66e874df7" in namespace "pods-3033" to be "running and ready" +Feb 25 11:29:23.842: INFO: Pod "server-envvars-cbbf4662-3990-4879-8518-62c66e874df7": Phase="Pending", Reason="", readiness=false. Elapsed: 5.284508ms +Feb 25 11:29:23.842: INFO: The phase of Pod server-envvars-cbbf4662-3990-4879-8518-62c66e874df7 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:29:25.849: INFO: Pod "server-envvars-cbbf4662-3990-4879-8518-62c66e874df7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012740956s +Feb 25 11:29:25.849: INFO: The phase of Pod server-envvars-cbbf4662-3990-4879-8518-62c66e874df7 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:29:27.849: INFO: Pod "server-envvars-cbbf4662-3990-4879-8518-62c66e874df7": Phase="Running", Reason="", readiness=true. Elapsed: 4.012676296s +Feb 25 11:29:27.849: INFO: The phase of Pod server-envvars-cbbf4662-3990-4879-8518-62c66e874df7 is Running (Ready = true) +Feb 25 11:29:27.849: INFO: Pod "server-envvars-cbbf4662-3990-4879-8518-62c66e874df7" satisfied condition "running and ready" +Feb 25 11:29:27.882: INFO: Waiting up to 5m0s for pod "client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad" in namespace "pods-3033" to be "Succeeded or Failed" +Feb 25 11:29:27.890: INFO: Pod "client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad": Phase="Pending", Reason="", readiness=false. Elapsed: 7.744158ms +Feb 25 11:29:29.897: INFO: Pod "client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014849024s +Feb 25 11:29:31.896: INFO: Pod "client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad": Phase="Pending", Reason="", readiness=false. Elapsed: 4.014327268s +Feb 25 11:29:33.897: INFO: Pod "client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.014926542s +STEP: Saw pod success 02/25/23 11:29:33.897 +Feb 25 11:29:33.897: INFO: Pod "client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad" satisfied condition "Succeeded or Failed" +Feb 25 11:29:33.909: INFO: Trying to get logs from node diek3dua9gay-2 pod client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad container env3cont: +STEP: delete the pod 02/25/23 11:29:33.936 +Feb 25 11:29:33.958: INFO: Waiting for pod client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad to disappear +Feb 25 11:29:33.964: INFO: Pod client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad no longer exists +[AfterEach] [sig-node] Pods + test/e2e/framework/framework.go:187 +Feb 25 11:29:33.965: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-3033" for this suite. 02/25/23 11:29:33.972 +{"msg":"PASSED [sig-node] Pods should contain environment variables for services [NodeConformance] [Conformance]","completed":308,"skipped":5961,"failed":0} +------------------------------ +• [SLOW TEST] [10.206 seconds] +[sig-node] Pods +test/e2e/common/node/framework.go:23 + should contain environment variables for services [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:443 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-node] Pods + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 11:29:23.778 + Feb 25 11:29:23.778: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename pods 02/25/23 11:29:23.78 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:29:23.809 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:29:23.813 + [BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 + [It] should contain environment variables for services [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:443 + Feb 25 11:29:23.837: INFO: Waiting up to 5m0s for pod "server-envvars-cbbf4662-3990-4879-8518-62c66e874df7" in namespace "pods-3033" to be "running and ready" + Feb 25 11:29:23.842: INFO: Pod "server-envvars-cbbf4662-3990-4879-8518-62c66e874df7": Phase="Pending", Reason="", readiness=false. Elapsed: 5.284508ms + Feb 25 11:29:23.842: INFO: The phase of Pod server-envvars-cbbf4662-3990-4879-8518-62c66e874df7 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:29:25.849: INFO: Pod "server-envvars-cbbf4662-3990-4879-8518-62c66e874df7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012740956s + Feb 25 11:29:25.849: INFO: The phase of Pod server-envvars-cbbf4662-3990-4879-8518-62c66e874df7 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:29:27.849: INFO: Pod "server-envvars-cbbf4662-3990-4879-8518-62c66e874df7": Phase="Running", Reason="", readiness=true. Elapsed: 4.012676296s + Feb 25 11:29:27.849: INFO: The phase of Pod server-envvars-cbbf4662-3990-4879-8518-62c66e874df7 is Running (Ready = true) + Feb 25 11:29:27.849: INFO: Pod "server-envvars-cbbf4662-3990-4879-8518-62c66e874df7" satisfied condition "running and ready" + Feb 25 11:29:27.882: INFO: Waiting up to 5m0s for pod "client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad" in namespace "pods-3033" to be "Succeeded or Failed" + Feb 25 11:29:27.890: INFO: Pod "client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad": Phase="Pending", Reason="", readiness=false. Elapsed: 7.744158ms + Feb 25 11:29:29.897: INFO: Pod "client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014849024s + Feb 25 11:29:31.896: INFO: Pod "client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad": Phase="Pending", Reason="", readiness=false. Elapsed: 4.014327268s + Feb 25 11:29:33.897: INFO: Pod "client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.014926542s + STEP: Saw pod success 02/25/23 11:29:33.897 + Feb 25 11:29:33.897: INFO: Pod "client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad" satisfied condition "Succeeded or Failed" + Feb 25 11:29:33.909: INFO: Trying to get logs from node diek3dua9gay-2 pod client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad container env3cont: + STEP: delete the pod 02/25/23 11:29:33.936 + Feb 25 11:29:33.958: INFO: Waiting for pod client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad to disappear + Feb 25 11:29:33.964: INFO: Pod client-envvars-877e9252-fe1d-44a4-bec6-c5ee03f670ad no longer exists + [AfterEach] [sig-node] Pods + test/e2e/framework/framework.go:187 + Feb 25 11:29:33.965: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "pods-3033" for this suite. 02/25/23 11:29:33.972 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSS +------------------------------ +[sig-cli] Kubectl client Kubectl version + should check is all data is printed [Conformance] + test/e2e/kubectl/kubectl.go:1683 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 11:29:33.986 +Feb 25 11:29:33.986: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 11:29:33.989 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:29:34.014 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:29:34.02 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 +[It] should check is all data is printed [Conformance] + test/e2e/kubectl/kubectl.go:1683 +Feb 25 11:29:34.026: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1202 version' +Feb 25 11:29:34.142: INFO: stderr: "WARNING: This version information is deprecated and will be replaced with the output from kubectl version --short. Use --output=yaml|json to get the full version.\n" +Feb 25 11:29:34.142: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"25\", GitVersion:\"v1.25.6\", GitCommit:\"ff2c119726cc1f8926fb0585c74b25921e866a28\", GitTreeState:\"clean\", BuildDate:\"2023-01-18T19:22:09Z\", GoVersion:\"go1.19.5\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nKustomize Version: v4.5.7\nServer Version: version.Info{Major:\"1\", Minor:\"25\", GitVersion:\"v1.25.6\", GitCommit:\"ff2c119726cc1f8926fb0585c74b25921e866a28\", GitTreeState:\"clean\", BuildDate:\"2023-01-18T19:15:26Z\", GoVersion:\"go1.19.5\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" +[AfterEach] [sig-cli] Kubectl client + test/e2e/framework/framework.go:187 +Feb 25 11:29:34.142: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-1202" for this suite. 02/25/23 11:29:34.156 +{"msg":"PASSED [sig-cli] Kubectl client Kubectl version should check is all data is printed [Conformance]","completed":309,"skipped":5976,"failed":0} +------------------------------ +• [0.183 seconds] +[sig-cli] Kubectl client +test/e2e/kubectl/framework.go:23 + Kubectl version + test/e2e/kubectl/kubectl.go:1677 + should check is all data is printed [Conformance] + test/e2e/kubectl/kubectl.go:1683 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-cli] Kubectl client + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 11:29:33.986 + Feb 25 11:29:33.986: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 11:29:33.989 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:29:34.014 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:29:34.02 + [BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 + [It] should check is all data is printed [Conformance] + test/e2e/kubectl/kubectl.go:1683 + Feb 25 11:29:34.026: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-1202 version' + Feb 25 11:29:34.142: INFO: stderr: "WARNING: This version information is deprecated and will be replaced with the output from kubectl version --short. Use --output=yaml|json to get the full version.\n" + Feb 25 11:29:34.142: INFO: stdout: "Client Version: version.Info{Major:\"1\", Minor:\"25\", GitVersion:\"v1.25.6\", GitCommit:\"ff2c119726cc1f8926fb0585c74b25921e866a28\", GitTreeState:\"clean\", BuildDate:\"2023-01-18T19:22:09Z\", GoVersion:\"go1.19.5\", Compiler:\"gc\", Platform:\"linux/amd64\"}\nKustomize Version: v4.5.7\nServer Version: version.Info{Major:\"1\", Minor:\"25\", GitVersion:\"v1.25.6\", GitCommit:\"ff2c119726cc1f8926fb0585c74b25921e866a28\", GitTreeState:\"clean\", BuildDate:\"2023-01-18T19:15:26Z\", GoVersion:\"go1.19.5\", Compiler:\"gc\", Platform:\"linux/amd64\"}\n" + [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 17:00:38.681: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-lifecycle-hook-4184" for this suite. 12/14/22 17:00:38.689 + Feb 25 11:29:34.142: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-1202" for this suite. 02/25/23 11:29:34.156 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSS +SSSS ------------------------------ -[sig-node] Pods - should be updated [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:343 -[BeforeEach] [sig-node] Pods +[sig-node] InitContainer [NodeConformance] + should not start app containers if init containers fail on a RestartAlways pod [Conformance] + test/e2e/common/node/init_container.go:333 +[BeforeEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:00:38.7 -Dec 14 17:00:38.700: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename pods 12/14/22 17:00:38.702 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:00:38.724 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:00:38.731 -[BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 -[It] should be updated [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:343 -STEP: creating the pod 12/14/22 17:00:38.736 -STEP: submitting the pod to kubernetes 12/14/22 17:00:38.736 -Dec 14 17:00:38.749: INFO: Waiting up to 5m0s for pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48" in namespace "pods-7593" to be "running and ready" -Dec 14 17:00:38.752: INFO: Pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48": Phase="Pending", Reason="", readiness=false. Elapsed: 3.176293ms -Dec 14 17:00:38.753: INFO: The phase of Pod pod-update-5828da98-6326-4b36-ac2b-bb3029872f48 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 17:00:40.760: INFO: Pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48": Phase="Running", Reason="", readiness=true. Elapsed: 2.010745582s -Dec 14 17:00:40.760: INFO: The phase of Pod pod-update-5828da98-6326-4b36-ac2b-bb3029872f48 is Running (Ready = true) -Dec 14 17:00:40.760: INFO: Pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48" satisfied condition "running and ready" -STEP: verifying the pod is in kubernetes 12/14/22 17:00:40.766 -STEP: updating the pod 12/14/22 17:00:40.772 -Dec 14 17:00:41.293: INFO: Successfully updated pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48" -Dec 14 17:00:41.293: INFO: Waiting up to 5m0s for pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48" in namespace "pods-7593" to be "running" -Dec 14 17:00:41.298: INFO: Pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48": Phase="Running", Reason="", readiness=true. Elapsed: 5.038602ms -Dec 14 17:00:41.298: INFO: Pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48" satisfied condition "running" -STEP: verifying the updated pod is in kubernetes 12/14/22 17:00:41.298 -Dec 14 17:00:41.303: INFO: Pod update OK -[AfterEach] [sig-node] Pods +STEP: Creating a kubernetes client 02/25/23 11:29:34.169 +Feb 25 11:29:34.169: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename init-container 02/25/23 11:29:34.172 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:29:34.202 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:29:34.206 +[BeforeEach] [sig-node] InitContainer [NodeConformance] + test/e2e/common/node/init_container.go:164 +[It] should not start app containers if init containers fail on a RestartAlways pod [Conformance] + test/e2e/common/node/init_container.go:333 +STEP: creating the pod 02/25/23 11:29:34.212 +Feb 25 11:29:34.212: INFO: PodSpec: initContainers in spec.initContainers +Feb 25 11:30:23.358: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-5eb9e8bc-eb7c-4d9a-a383-0e1ffddc463c", GenerateName:"", Namespace:"init-container-6591", SelfLink:"", UID:"e656b1fe-802c-4c89-bec6-c8de4d5ae109", ResourceVersion:"43347", Generation:0, CreationTimestamp:time.Date(2023, time.February, 25, 11, 29, 34, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"212583583"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:"e2e.test", Operation:"Update", APIVersion:"v1", Time:time.Date(2023, time.February, 25, 11, 29, 34, 0, time.Local), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc000d2c1f8), Subresource:""}, v1.ManagedFieldsEntry{Manager:"kubelet", Operation:"Update", APIVersion:"v1", Time:time.Date(2023, time.February, 25, 11, 30, 23, 0, time.Local), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc000d2c240), Subresource:"status"}}}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"kube-api-access-6tc4c", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(nil), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(0xc0042afe20), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil), Ephemeral:(*v1.EphemeralVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-api-access-6tc4c", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-api-access-6tc4c", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"registry.k8s.io/pause:3.8", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-api-access-6tc4c", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, EphemeralContainers:[]v1.EphemeralContainer(nil), RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc002b50f50), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"diek3dua9gay-3", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc0022b6700), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc002b50fe0)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc002b51000)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc002b51008), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc002b5100c), PreemptionPolicy:(*v1.PreemptionPolicy)(0xc003d0c7c0), Overhead:v1.ResourceList(nil), TopologySpreadConstraints:[]v1.TopologySpreadConstraint(nil), SetHostnameAsFQDN:(*bool)(nil), OS:(*v1.PodOS)(nil), HostUsers:(*bool)(nil)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2023, time.February, 25, 11, 29, 34, 0, time.Local), Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2023, time.February, 25, 11, 29, 34, 0, time.Local), Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2023, time.February, 25, 11, 29, 34, 0, time.Local), Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2023, time.February, 25, 11, 29, 34, 0, time.Local), Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"192.168.121.44", PodIP:"10.233.66.252", PodIPs:[]v1.PodIP{v1.PodIP{IP:"10.233.66.252"}}, StartTime:time.Date(2023, time.February, 25, 11, 29, 34, 0, time.Local), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc0022b67e0)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc0022b6850)}, Ready:false, RestartCount:3, Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", ImageID:"registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf", ContainerID:"cri-o://a6ee43cdddd557b59fa29c019e2993791e8c2615ae568152d50b7f323b327c19", Started:(*bool)(nil)}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc0042afea0), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", ImageID:"", ContainerID:"", Started:(*bool)(nil)}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc0042afe80), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"registry.k8s.io/pause:3.8", ImageID:"", ContainerID:"", Started:(*bool)(0xc002b510a4)}}, QOSClass:"Burstable", EphemeralContainerStatuses:[]v1.ContainerStatus(nil)}} +[AfterEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:187 -Dec 14 17:00:41.303: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pods-7593" for this suite. 12/14/22 17:00:41.312 -{"msg":"PASSED [sig-node] Pods should be updated [NodeConformance] [Conformance]","completed":312,"skipped":5911,"failed":0} +Feb 25 11:30:23.367: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "init-container-6591" for this suite. 02/25/23 11:30:23.378 +{"msg":"PASSED [sig-node] InitContainer [NodeConformance] should not start app containers if init containers fail on a RestartAlways pod [Conformance]","completed":310,"skipped":5980,"failed":0} ------------------------------ -• [2.622 seconds] -[sig-node] Pods +• [SLOW TEST] [49.229 seconds] +[sig-node] InitContainer [NodeConformance] test/e2e/common/node/framework.go:23 - should be updated [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:343 + should not start app containers if init containers fail on a RestartAlways pod [Conformance] + test/e2e/common/node/init_container.go:333 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Pods + [BeforeEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:00:38.7 - Dec 14 17:00:38.700: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename pods 12/14/22 17:00:38.702 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:00:38.724 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:00:38.731 - [BeforeEach] [sig-node] Pods - test/e2e/common/node/pods.go:193 - [It] should be updated [NodeConformance] [Conformance] - test/e2e/common/node/pods.go:343 - STEP: creating the pod 12/14/22 17:00:38.736 - STEP: submitting the pod to kubernetes 12/14/22 17:00:38.736 - Dec 14 17:00:38.749: INFO: Waiting up to 5m0s for pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48" in namespace "pods-7593" to be "running and ready" - Dec 14 17:00:38.752: INFO: Pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48": Phase="Pending", Reason="", readiness=false. Elapsed: 3.176293ms - Dec 14 17:00:38.753: INFO: The phase of Pod pod-update-5828da98-6326-4b36-ac2b-bb3029872f48 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 17:00:40.760: INFO: Pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48": Phase="Running", Reason="", readiness=true. Elapsed: 2.010745582s - Dec 14 17:00:40.760: INFO: The phase of Pod pod-update-5828da98-6326-4b36-ac2b-bb3029872f48 is Running (Ready = true) - Dec 14 17:00:40.760: INFO: Pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48" satisfied condition "running and ready" - STEP: verifying the pod is in kubernetes 12/14/22 17:00:40.766 - STEP: updating the pod 12/14/22 17:00:40.772 - Dec 14 17:00:41.293: INFO: Successfully updated pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48" - Dec 14 17:00:41.293: INFO: Waiting up to 5m0s for pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48" in namespace "pods-7593" to be "running" - Dec 14 17:00:41.298: INFO: Pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48": Phase="Running", Reason="", readiness=true. Elapsed: 5.038602ms - Dec 14 17:00:41.298: INFO: Pod "pod-update-5828da98-6326-4b36-ac2b-bb3029872f48" satisfied condition "running" - STEP: verifying the updated pod is in kubernetes 12/14/22 17:00:41.298 - Dec 14 17:00:41.303: INFO: Pod update OK - [AfterEach] [sig-node] Pods + STEP: Creating a kubernetes client 02/25/23 11:29:34.169 + Feb 25 11:29:34.169: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename init-container 02/25/23 11:29:34.172 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:29:34.202 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:29:34.206 + [BeforeEach] [sig-node] InitContainer [NodeConformance] + test/e2e/common/node/init_container.go:164 + [It] should not start app containers if init containers fail on a RestartAlways pod [Conformance] + test/e2e/common/node/init_container.go:333 + STEP: creating the pod 02/25/23 11:29:34.212 + Feb 25 11:29:34.212: INFO: PodSpec: initContainers in spec.initContainers + Feb 25 11:30:23.358: INFO: init container has failed twice: &v1.Pod{TypeMeta:v1.TypeMeta{Kind:"", APIVersion:""}, ObjectMeta:v1.ObjectMeta{Name:"pod-init-5eb9e8bc-eb7c-4d9a-a383-0e1ffddc463c", GenerateName:"", Namespace:"init-container-6591", SelfLink:"", UID:"e656b1fe-802c-4c89-bec6-c8de4d5ae109", ResourceVersion:"43347", Generation:0, CreationTimestamp:time.Date(2023, time.February, 25, 11, 29, 34, 0, time.Local), DeletionTimestamp:, DeletionGracePeriodSeconds:(*int64)(nil), Labels:map[string]string{"name":"foo", "time":"212583583"}, Annotations:map[string]string(nil), OwnerReferences:[]v1.OwnerReference(nil), Finalizers:[]string(nil), ManagedFields:[]v1.ManagedFieldsEntry{v1.ManagedFieldsEntry{Manager:"e2e.test", Operation:"Update", APIVersion:"v1", Time:time.Date(2023, time.February, 25, 11, 29, 34, 0, time.Local), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc000d2c1f8), Subresource:""}, v1.ManagedFieldsEntry{Manager:"kubelet", Operation:"Update", APIVersion:"v1", Time:time.Date(2023, time.February, 25, 11, 30, 23, 0, time.Local), FieldsType:"FieldsV1", FieldsV1:(*v1.FieldsV1)(0xc000d2c240), Subresource:"status"}}}, Spec:v1.PodSpec{Volumes:[]v1.Volume{v1.Volume{Name:"kube-api-access-6tc4c", VolumeSource:v1.VolumeSource{HostPath:(*v1.HostPathVolumeSource)(nil), EmptyDir:(*v1.EmptyDirVolumeSource)(nil), GCEPersistentDisk:(*v1.GCEPersistentDiskVolumeSource)(nil), AWSElasticBlockStore:(*v1.AWSElasticBlockStoreVolumeSource)(nil), GitRepo:(*v1.GitRepoVolumeSource)(nil), Secret:(*v1.SecretVolumeSource)(nil), NFS:(*v1.NFSVolumeSource)(nil), ISCSI:(*v1.ISCSIVolumeSource)(nil), Glusterfs:(*v1.GlusterfsVolumeSource)(nil), PersistentVolumeClaim:(*v1.PersistentVolumeClaimVolumeSource)(nil), RBD:(*v1.RBDVolumeSource)(nil), FlexVolume:(*v1.FlexVolumeSource)(nil), Cinder:(*v1.CinderVolumeSource)(nil), CephFS:(*v1.CephFSVolumeSource)(nil), Flocker:(*v1.FlockerVolumeSource)(nil), DownwardAPI:(*v1.DownwardAPIVolumeSource)(nil), FC:(*v1.FCVolumeSource)(nil), AzureFile:(*v1.AzureFileVolumeSource)(nil), ConfigMap:(*v1.ConfigMapVolumeSource)(nil), VsphereVolume:(*v1.VsphereVirtualDiskVolumeSource)(nil), Quobyte:(*v1.QuobyteVolumeSource)(nil), AzureDisk:(*v1.AzureDiskVolumeSource)(nil), PhotonPersistentDisk:(*v1.PhotonPersistentDiskVolumeSource)(nil), Projected:(*v1.ProjectedVolumeSource)(0xc0042afe20), PortworxVolume:(*v1.PortworxVolumeSource)(nil), ScaleIO:(*v1.ScaleIOVolumeSource)(nil), StorageOS:(*v1.StorageOSVolumeSource)(nil), CSI:(*v1.CSIVolumeSource)(nil), Ephemeral:(*v1.EphemeralVolumeSource)(nil)}}}, InitContainers:[]v1.Container{v1.Container{Name:"init1", Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", Command:[]string{"/bin/false"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-api-access-6tc4c", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}, v1.Container{Name:"init2", Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", Command:[]string{"/bin/true"}, Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList(nil), Requests:v1.ResourceList(nil)}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-api-access-6tc4c", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, Containers:[]v1.Container{v1.Container{Name:"run1", Image:"registry.k8s.io/pause:3.8", Command:[]string(nil), Args:[]string(nil), WorkingDir:"", Ports:[]v1.ContainerPort(nil), EnvFrom:[]v1.EnvFromSource(nil), Env:[]v1.EnvVar(nil), Resources:v1.ResourceRequirements{Limits:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}, Requests:v1.ResourceList{"cpu":resource.Quantity{i:resource.int64Amount{value:100, scale:-3}, d:resource.infDecAmount{Dec:(*inf.Dec)(nil)}, s:"100m", Format:"DecimalSI"}}}, VolumeMounts:[]v1.VolumeMount{v1.VolumeMount{Name:"kube-api-access-6tc4c", ReadOnly:true, MountPath:"/var/run/secrets/kubernetes.io/serviceaccount", SubPath:"", MountPropagation:(*v1.MountPropagationMode)(nil), SubPathExpr:""}}, VolumeDevices:[]v1.VolumeDevice(nil), LivenessProbe:(*v1.Probe)(nil), ReadinessProbe:(*v1.Probe)(nil), StartupProbe:(*v1.Probe)(nil), Lifecycle:(*v1.Lifecycle)(nil), TerminationMessagePath:"/dev/termination-log", TerminationMessagePolicy:"File", ImagePullPolicy:"IfNotPresent", SecurityContext:(*v1.SecurityContext)(nil), Stdin:false, StdinOnce:false, TTY:false}}, EphemeralContainers:[]v1.EphemeralContainer(nil), RestartPolicy:"Always", TerminationGracePeriodSeconds:(*int64)(0xc002b50f50), ActiveDeadlineSeconds:(*int64)(nil), DNSPolicy:"ClusterFirst", NodeSelector:map[string]string(nil), ServiceAccountName:"default", DeprecatedServiceAccount:"default", AutomountServiceAccountToken:(*bool)(nil), NodeName:"diek3dua9gay-3", HostNetwork:false, HostPID:false, HostIPC:false, ShareProcessNamespace:(*bool)(nil), SecurityContext:(*v1.PodSecurityContext)(0xc0022b6700), ImagePullSecrets:[]v1.LocalObjectReference(nil), Hostname:"", Subdomain:"", Affinity:(*v1.Affinity)(nil), SchedulerName:"default-scheduler", Tolerations:[]v1.Toleration{v1.Toleration{Key:"node.kubernetes.io/not-ready", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc002b50fe0)}, v1.Toleration{Key:"node.kubernetes.io/unreachable", Operator:"Exists", Value:"", Effect:"NoExecute", TolerationSeconds:(*int64)(0xc002b51000)}}, HostAliases:[]v1.HostAlias(nil), PriorityClassName:"", Priority:(*int32)(0xc002b51008), DNSConfig:(*v1.PodDNSConfig)(nil), ReadinessGates:[]v1.PodReadinessGate(nil), RuntimeClassName:(*string)(nil), EnableServiceLinks:(*bool)(0xc002b5100c), PreemptionPolicy:(*v1.PreemptionPolicy)(0xc003d0c7c0), Overhead:v1.ResourceList(nil), TopologySpreadConstraints:[]v1.TopologySpreadConstraint(nil), SetHostnameAsFQDN:(*bool)(nil), OS:(*v1.PodOS)(nil), HostUsers:(*bool)(nil)}, Status:v1.PodStatus{Phase:"Pending", Conditions:[]v1.PodCondition{v1.PodCondition{Type:"Initialized", Status:"False", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2023, time.February, 25, 11, 29, 34, 0, time.Local), Reason:"ContainersNotInitialized", Message:"containers with incomplete status: [init1 init2]"}, v1.PodCondition{Type:"Ready", Status:"False", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2023, time.February, 25, 11, 29, 34, 0, time.Local), Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"ContainersReady", Status:"False", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2023, time.February, 25, 11, 29, 34, 0, time.Local), Reason:"ContainersNotReady", Message:"containers with unready status: [run1]"}, v1.PodCondition{Type:"PodScheduled", Status:"True", LastProbeTime:time.Date(1, time.January, 1, 0, 0, 0, 0, time.UTC), LastTransitionTime:time.Date(2023, time.February, 25, 11, 29, 34, 0, time.Local), Reason:"", Message:""}}, Message:"", Reason:"", NominatedNodeName:"", HostIP:"192.168.121.44", PodIP:"10.233.66.252", PodIPs:[]v1.PodIP{v1.PodIP{IP:"10.233.66.252"}}, StartTime:time.Date(2023, time.February, 25, 11, 29, 34, 0, time.Local), InitContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"init1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc0022b67e0)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(0xc0022b6850)}, Ready:false, RestartCount:3, Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", ImageID:"registry.k8s.io/e2e-test-images/busybox@sha256:c318242786b139d18676b1c09a0ad7f15fc17f8f16a5b2e625cd0dc8c9703daf", ContainerID:"cri-o://a6ee43cdddd557b59fa29c019e2993791e8c2615ae568152d50b7f323b327c19", Started:(*bool)(nil)}, v1.ContainerStatus{Name:"init2", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc0042afea0), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"registry.k8s.io/e2e-test-images/busybox:1.29-2", ImageID:"", ContainerID:"", Started:(*bool)(nil)}}, ContainerStatuses:[]v1.ContainerStatus{v1.ContainerStatus{Name:"run1", State:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(0xc0042afe80), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, LastTerminationState:v1.ContainerState{Waiting:(*v1.ContainerStateWaiting)(nil), Running:(*v1.ContainerStateRunning)(nil), Terminated:(*v1.ContainerStateTerminated)(nil)}, Ready:false, RestartCount:0, Image:"registry.k8s.io/pause:3.8", ImageID:"", ContainerID:"", Started:(*bool)(0xc002b510a4)}}, QOSClass:"Burstable", EphemeralContainerStatuses:[]v1.ContainerStatus(nil)}} + [AfterEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:187 - Dec 14 17:00:41.303: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "pods-7593" for this suite. 12/14/22 17:00:41.312 + Feb 25 11:30:23.367: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "init-container-6591" for this suite. 02/25/23 11:30:23.378 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] server version - should find the server version [Conformance] - test/e2e/apimachinery/server_version.go:39 -[BeforeEach] [sig-api-machinery] server version +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] + works for multiple CRDs of same group and version but different kinds [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:356 +[BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:00:41.326 -Dec 14 17:00:41.327: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename server-version 12/14/22 17:00:41.33 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:00:41.357 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:00:41.361 -[It] should find the server version [Conformance] - test/e2e/apimachinery/server_version.go:39 -STEP: Request ServerVersion 12/14/22 17:00:41.366 -STEP: Confirm major version 12/14/22 17:00:41.37 -Dec 14 17:00:41.370: INFO: Major version: 1 -STEP: Confirm minor version 12/14/22 17:00:41.37 -Dec 14 17:00:41.370: INFO: cleanMinorVersion: 25 -Dec 14 17:00:41.370: INFO: Minor version: 25 -[AfterEach] [sig-api-machinery] server version +STEP: Creating a kubernetes client 02/25/23 11:30:23.411 +Feb 25 11:30:23.412: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 11:30:23.415 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:30:23.444 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:30:23.449 +[It] works for multiple CRDs of same group and version but different kinds [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:356 +STEP: CRs in the same group and version but different kinds (two CRDs) show up in OpenAPI documentation 02/25/23 11:30:23.454 +Feb 25 11:30:23.455: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 11:30:29.610: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +[AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 17:00:41.370: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "server-version-3830" for this suite. 12/14/22 17:00:41.375 -{"msg":"PASSED [sig-api-machinery] server version should find the server version [Conformance]","completed":313,"skipped":5921,"failed":0} +Feb 25 11:30:48.050: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "crd-publish-openapi-3944" for this suite. 02/25/23 11:30:48.064 +{"msg":"PASSED [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] works for multiple CRDs of same group and version but different kinds [Conformance]","completed":311,"skipped":6017,"failed":0} ------------------------------ -• [0.068 seconds] -[sig-api-machinery] server version +• [SLOW TEST] [24.664 seconds] +[sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/apimachinery/framework.go:23 - should find the server version [Conformance] - test/e2e/apimachinery/server_version.go:39 + works for multiple CRDs of same group and version but different kinds [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:356 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] server version + [BeforeEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:00:41.326 - Dec 14 17:00:41.327: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename server-version 12/14/22 17:00:41.33 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:00:41.357 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:00:41.361 - [It] should find the server version [Conformance] - test/e2e/apimachinery/server_version.go:39 - STEP: Request ServerVersion 12/14/22 17:00:41.366 - STEP: Confirm major version 12/14/22 17:00:41.37 - Dec 14 17:00:41.370: INFO: Major version: 1 - STEP: Confirm minor version 12/14/22 17:00:41.37 - Dec 14 17:00:41.370: INFO: cleanMinorVersion: 25 - Dec 14 17:00:41.370: INFO: Minor version: 25 - [AfterEach] [sig-api-machinery] server version + STEP: Creating a kubernetes client 02/25/23 11:30:23.411 + Feb 25 11:30:23.412: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename crd-publish-openapi 02/25/23 11:30:23.415 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:30:23.444 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:30:23.449 + [It] works for multiple CRDs of same group and version but different kinds [Conformance] + test/e2e/apimachinery/crd_publish_openapi.go:356 + STEP: CRs in the same group and version but different kinds (two CRDs) show up in OpenAPI documentation 02/25/23 11:30:23.454 + Feb 25 11:30:23.455: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 11:30:29.610: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + [AfterEach] [sig-api-machinery] CustomResourcePublishOpenAPI [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 17:00:41.370: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "server-version-3830" for this suite. 12/14/22 17:00:41.375 + Feb 25 11:30:48.050: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "crd-publish-openapi-3944" for this suite. 02/25/23 11:30:48.064 << End Captured GinkgoWriter Output ------------------------------ S ------------------------------ -[sig-node] Probing container - should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:211 -[BeforeEach] [sig-node] Probing container +[sig-node] ConfigMap + should run through a ConfigMap lifecycle [Conformance] + test/e2e/common/node/configmap.go:168 +[BeforeEach] [sig-node] ConfigMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:00:41.399 -Dec 14 17:00:41.400: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-probe 12/14/22 17:00:41.402 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:00:41.427 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:00:41.431 -[BeforeEach] [sig-node] Probing container - test/e2e/common/node/container_probe.go:59 -[It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:211 -STEP: Creating pod test-webserver-e4cc8d8d-c3ba-4187-a538-e263e51f32af in namespace container-probe-7546 12/14/22 17:00:41.436 -Dec 14 17:00:41.449: INFO: Waiting up to 5m0s for pod "test-webserver-e4cc8d8d-c3ba-4187-a538-e263e51f32af" in namespace "container-probe-7546" to be "not pending" -Dec 14 17:00:41.455: INFO: Pod "test-webserver-e4cc8d8d-c3ba-4187-a538-e263e51f32af": Phase="Pending", Reason="", readiness=false. Elapsed: 5.513629ms -Dec 14 17:00:43.463: INFO: Pod "test-webserver-e4cc8d8d-c3ba-4187-a538-e263e51f32af": Phase="Running", Reason="", readiness=true. Elapsed: 2.013676689s -Dec 14 17:00:43.463: INFO: Pod "test-webserver-e4cc8d8d-c3ba-4187-a538-e263e51f32af" satisfied condition "not pending" -Dec 14 17:00:43.463: INFO: Started pod test-webserver-e4cc8d8d-c3ba-4187-a538-e263e51f32af in namespace container-probe-7546 -STEP: checking the pod's current state and verifying that restartCount is present 12/14/22 17:00:43.463 -Dec 14 17:00:43.469: INFO: Initial restart count of pod test-webserver-e4cc8d8d-c3ba-4187-a538-e263e51f32af is 0 -STEP: deleting the pod 12/14/22 17:04:44.459 -[AfterEach] [sig-node] Probing container +STEP: Creating a kubernetes client 02/25/23 11:30:48.077 +Feb 25 11:30:48.077: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename configmap 02/25/23 11:30:48.079 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:30:48.115 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:30:48.119 +[It] should run through a ConfigMap lifecycle [Conformance] + test/e2e/common/node/configmap.go:168 +STEP: creating a ConfigMap 02/25/23 11:30:48.133 +STEP: fetching the ConfigMap 02/25/23 11:30:48.143 +STEP: patching the ConfigMap 02/25/23 11:30:48.147 +STEP: listing all ConfigMaps in all namespaces with a label selector 02/25/23 11:30:48.157 +STEP: deleting the ConfigMap by collection with a label selector 02/25/23 11:30:48.164 +STEP: listing all ConfigMaps in test namespace 02/25/23 11:30:48.177 +[AfterEach] [sig-node] ConfigMap test/e2e/framework/framework.go:187 -Dec 14 17:04:44.487: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-probe-7546" for this suite. 12/14/22 17:04:44.497 -{"msg":"PASSED [sig-node] Probing container should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","completed":314,"skipped":5922,"failed":0} +Feb 25 11:30:48.182: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-105" for this suite. 02/25/23 11:30:48.188 +{"msg":"PASSED [sig-node] ConfigMap should run through a ConfigMap lifecycle [Conformance]","completed":312,"skipped":6018,"failed":0} ------------------------------ -• [SLOW TEST] [243.120 seconds] -[sig-node] Probing container +• [0.120 seconds] +[sig-node] ConfigMap test/e2e/common/node/framework.go:23 - should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:211 + should run through a ConfigMap lifecycle [Conformance] + test/e2e/common/node/configmap.go:168 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Probing container + [BeforeEach] [sig-node] ConfigMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:00:41.399 - Dec 14 17:00:41.400: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-probe 12/14/22 17:00:41.402 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:00:41.427 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:00:41.431 - [BeforeEach] [sig-node] Probing container - test/e2e/common/node/container_probe.go:59 - [It] should *not* be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:211 - STEP: Creating pod test-webserver-e4cc8d8d-c3ba-4187-a538-e263e51f32af in namespace container-probe-7546 12/14/22 17:00:41.436 - Dec 14 17:00:41.449: INFO: Waiting up to 5m0s for pod "test-webserver-e4cc8d8d-c3ba-4187-a538-e263e51f32af" in namespace "container-probe-7546" to be "not pending" - Dec 14 17:00:41.455: INFO: Pod "test-webserver-e4cc8d8d-c3ba-4187-a538-e263e51f32af": Phase="Pending", Reason="", readiness=false. Elapsed: 5.513629ms - Dec 14 17:00:43.463: INFO: Pod "test-webserver-e4cc8d8d-c3ba-4187-a538-e263e51f32af": Phase="Running", Reason="", readiness=true. Elapsed: 2.013676689s - Dec 14 17:00:43.463: INFO: Pod "test-webserver-e4cc8d8d-c3ba-4187-a538-e263e51f32af" satisfied condition "not pending" - Dec 14 17:00:43.463: INFO: Started pod test-webserver-e4cc8d8d-c3ba-4187-a538-e263e51f32af in namespace container-probe-7546 - STEP: checking the pod's current state and verifying that restartCount is present 12/14/22 17:00:43.463 - Dec 14 17:00:43.469: INFO: Initial restart count of pod test-webserver-e4cc8d8d-c3ba-4187-a538-e263e51f32af is 0 - STEP: deleting the pod 12/14/22 17:04:44.459 - [AfterEach] [sig-node] Probing container + STEP: Creating a kubernetes client 02/25/23 11:30:48.077 + Feb 25 11:30:48.077: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename configmap 02/25/23 11:30:48.079 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:30:48.115 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:30:48.119 + [It] should run through a ConfigMap lifecycle [Conformance] + test/e2e/common/node/configmap.go:168 + STEP: creating a ConfigMap 02/25/23 11:30:48.133 + STEP: fetching the ConfigMap 02/25/23 11:30:48.143 + STEP: patching the ConfigMap 02/25/23 11:30:48.147 + STEP: listing all ConfigMaps in all namespaces with a label selector 02/25/23 11:30:48.157 + STEP: deleting the ConfigMap by collection with a label selector 02/25/23 11:30:48.164 + STEP: listing all ConfigMaps in test namespace 02/25/23 11:30:48.177 + [AfterEach] [sig-node] ConfigMap test/e2e/framework/framework.go:187 - Dec 14 17:04:44.487: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-probe-7546" for this suite. 12/14/22 17:04:44.497 + Feb 25 11:30:48.182: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "configmap-105" for this suite. 02/25/23 11:30:48.188 << End Captured GinkgoWriter Output ------------------------------ -SSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-cli] Kubectl client Kubectl logs - should be able to retrieve and filter logs [Conformance] - test/e2e/kubectl/kubectl.go:1590 -[BeforeEach] [sig-cli] Kubectl client +[sig-api-machinery] Servers with support for Table transformation + should return a 406 for a backend which does not implement metadata [Conformance] + test/e2e/apimachinery/table_conversion.go:154 +[BeforeEach] [sig-api-machinery] Servers with support for Table transformation test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:04:44.52 -Dec 14 17:04:44.520: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 17:04:44.522 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:04:44.558 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:04:44.561 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 -[BeforeEach] Kubectl logs - test/e2e/kubectl/kubectl.go:1570 -STEP: creating an pod 12/14/22 17:04:44.564 -Dec 14 17:04:44.567: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 run logs-generator --image=registry.k8s.io/e2e-test-images/agnhost:2.40 --restart=Never --pod-running-timeout=2m0s -- logs-generator --log-lines-total 100 --run-duration 20s' -Dec 14 17:04:44.736: INFO: stderr: "" -Dec 14 17:04:44.737: INFO: stdout: "pod/logs-generator created\n" -[It] should be able to retrieve and filter logs [Conformance] - test/e2e/kubectl/kubectl.go:1590 -STEP: Waiting for log generator to start. 12/14/22 17:04:44.737 -Dec 14 17:04:44.737: INFO: Waiting up to 5m0s for 1 pods to be running and ready, or succeeded: [logs-generator] -Dec 14 17:04:44.737: INFO: Waiting up to 5m0s for pod "logs-generator" in namespace "kubectl-2193" to be "running and ready, or succeeded" -Dec 14 17:04:44.747: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 9.956395ms -Dec 14 17:04:44.747: INFO: Error evaluating pod condition running and ready, or succeeded: want pod 'logs-generator' on 'kiedai4iakae-3' to be 'Running' but was 'Pending' -Dec 14 17:04:46.752: INFO: Pod "logs-generator": Phase="Running", Reason="", readiness=true. Elapsed: 2.014771871s -Dec 14 17:04:46.752: INFO: Pod "logs-generator" satisfied condition "running and ready, or succeeded" -Dec 14 17:04:46.752: INFO: Wanted all 1 pods to be running and ready, or succeeded. Result: true. Pods: [logs-generator] -STEP: checking for a matching strings 12/14/22 17:04:46.752 -Dec 14 17:04:46.752: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 logs logs-generator logs-generator' -Dec 14 17:04:46.919: INFO: stderr: "" -Dec 14 17:04:46.919: INFO: stdout: "I1214 17:04:45.628421 1 logs_generator.go:76] 0 GET /api/v1/namespaces/default/pods/lcq 517\nI1214 17:04:45.828378 1 logs_generator.go:76] 1 GET /api/v1/namespaces/default/pods/sp2 467\nI1214 17:04:46.028473 1 logs_generator.go:76] 2 POST /api/v1/namespaces/kube-system/pods/6jsk 469\nI1214 17:04:46.228919 1 logs_generator.go:76] 3 GET /api/v1/namespaces/default/pods/jrm 522\nI1214 17:04:46.429442 1 logs_generator.go:76] 4 POST /api/v1/namespaces/default/pods/c2w8 581\nI1214 17:04:46.629167 1 logs_generator.go:76] 5 POST /api/v1/namespaces/kube-system/pods/m6s 439\nI1214 17:04:46.829530 1 logs_generator.go:76] 6 GET /api/v1/namespaces/ns/pods/grxd 358\n" -STEP: limiting log lines 12/14/22 17:04:46.92 -Dec 14 17:04:46.920: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 logs logs-generator logs-generator --tail=1' -Dec 14 17:04:47.087: INFO: stderr: "" -Dec 14 17:04:47.087: INFO: stdout: "I1214 17:04:47.028916 1 logs_generator.go:76] 7 GET /api/v1/namespaces/kube-system/pods/p4ws 494\n" -Dec 14 17:04:47.087: INFO: got output "I1214 17:04:47.028916 1 logs_generator.go:76] 7 GET /api/v1/namespaces/kube-system/pods/p4ws 494\n" -STEP: limiting log bytes 12/14/22 17:04:47.087 -Dec 14 17:04:47.087: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 logs logs-generator logs-generator --limit-bytes=1' -Dec 14 17:04:47.366: INFO: stderr: "" -Dec 14 17:04:47.366: INFO: stdout: "I" -Dec 14 17:04:47.366: INFO: got output "I" -STEP: exposing timestamps 12/14/22 17:04:47.366 -Dec 14 17:04:47.367: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 logs logs-generator logs-generator --tail=1 --timestamps' -Dec 14 17:04:47.498: INFO: stderr: "" -Dec 14 17:04:47.498: INFO: stdout: "2022-12-14T17:04:47.428573564Z I1214 17:04:47.428503 1 logs_generator.go:76] 9 GET /api/v1/namespaces/ns/pods/6fl 376\n" -Dec 14 17:04:47.498: INFO: got output "2022-12-14T17:04:47.428573564Z I1214 17:04:47.428503 1 logs_generator.go:76] 9 GET /api/v1/namespaces/ns/pods/6fl 376\n" -STEP: restricting to a time range 12/14/22 17:04:47.498 -Dec 14 17:04:49.998: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 logs logs-generator logs-generator --since=1s' -Dec 14 17:04:50.239: INFO: stderr: "" -Dec 14 17:04:50.240: INFO: stdout: "I1214 17:04:49.429045 1 logs_generator.go:76] 19 POST /api/v1/namespaces/ns/pods/gqk 358\nI1214 17:04:49.628413 1 logs_generator.go:76] 20 GET /api/v1/namespaces/default/pods/h6f 478\nI1214 17:04:49.828797 1 logs_generator.go:76] 21 POST /api/v1/namespaces/default/pods/kwz 394\nI1214 17:04:50.029512 1 logs_generator.go:76] 22 POST /api/v1/namespaces/default/pods/6qlg 384\nI1214 17:04:50.228738 1 logs_generator.go:76] 23 PUT /api/v1/namespaces/default/pods/pnf 459\n" -Dec 14 17:04:50.240: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 logs logs-generator logs-generator --since=24h' -Dec 14 17:04:50.404: INFO: stderr: "" -Dec 14 17:04:50.404: INFO: stdout: "I1214 17:04:45.628421 1 logs_generator.go:76] 0 GET /api/v1/namespaces/default/pods/lcq 517\nI1214 17:04:45.828378 1 logs_generator.go:76] 1 GET /api/v1/namespaces/default/pods/sp2 467\nI1214 17:04:46.028473 1 logs_generator.go:76] 2 POST /api/v1/namespaces/kube-system/pods/6jsk 469\nI1214 17:04:46.228919 1 logs_generator.go:76] 3 GET /api/v1/namespaces/default/pods/jrm 522\nI1214 17:04:46.429442 1 logs_generator.go:76] 4 POST /api/v1/namespaces/default/pods/c2w8 581\nI1214 17:04:46.629167 1 logs_generator.go:76] 5 POST /api/v1/namespaces/kube-system/pods/m6s 439\nI1214 17:04:46.829530 1 logs_generator.go:76] 6 GET /api/v1/namespaces/ns/pods/grxd 358\nI1214 17:04:47.028916 1 logs_generator.go:76] 7 GET /api/v1/namespaces/kube-system/pods/p4ws 494\nI1214 17:04:47.228849 1 logs_generator.go:76] 8 POST /api/v1/namespaces/ns/pods/mhq 474\nI1214 17:04:47.428503 1 logs_generator.go:76] 9 GET /api/v1/namespaces/ns/pods/6fl 376\nI1214 17:04:47.629133 1 logs_generator.go:76] 10 GET /api/v1/namespaces/default/pods/lh6 454\nI1214 17:04:47.828368 1 logs_generator.go:76] 11 PUT /api/v1/namespaces/default/pods/k24n 382\nI1214 17:04:48.028922 1 logs_generator.go:76] 12 GET /api/v1/namespaces/ns/pods/7hzq 252\nI1214 17:04:48.228318 1 logs_generator.go:76] 13 GET /api/v1/namespaces/default/pods/n9jf 415\nI1214 17:04:48.429076 1 logs_generator.go:76] 14 GET /api/v1/namespaces/kube-system/pods/9df 378\nI1214 17:04:48.628403 1 logs_generator.go:76] 15 PUT /api/v1/namespaces/default/pods/sj7w 521\nI1214 17:04:48.828790 1 logs_generator.go:76] 16 GET /api/v1/namespaces/default/pods/vvnl 542\nI1214 17:04:49.029166 1 logs_generator.go:76] 17 PUT /api/v1/namespaces/ns/pods/hdmn 541\nI1214 17:04:49.228478 1 logs_generator.go:76] 18 PUT /api/v1/namespaces/kube-system/pods/ct6 414\nI1214 17:04:49.429045 1 logs_generator.go:76] 19 POST /api/v1/namespaces/ns/pods/gqk 358\nI1214 17:04:49.628413 1 logs_generator.go:76] 20 GET /api/v1/namespaces/default/pods/h6f 478\nI1214 17:04:49.828797 1 logs_generator.go:76] 21 POST /api/v1/namespaces/default/pods/kwz 394\nI1214 17:04:50.029512 1 logs_generator.go:76] 22 POST /api/v1/namespaces/default/pods/6qlg 384\nI1214 17:04:50.228738 1 logs_generator.go:76] 23 PUT /api/v1/namespaces/default/pods/pnf 459\n" -[AfterEach] Kubectl logs - test/e2e/kubectl/kubectl.go:1575 -Dec 14 17:04:50.405: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 delete pod logs-generator' -Dec 14 17:04:51.553: INFO: stderr: "" -Dec 14 17:04:51.553: INFO: stdout: "pod \"logs-generator\" deleted\n" -[AfterEach] [sig-cli] Kubectl client +STEP: Creating a kubernetes client 02/25/23 11:30:48.203 +Feb 25 11:30:48.203: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename tables 02/25/23 11:30:48.207 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:30:48.234 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:30:48.238 +[BeforeEach] [sig-api-machinery] Servers with support for Table transformation + test/e2e/apimachinery/table_conversion.go:49 +[It] should return a 406 for a backend which does not implement metadata [Conformance] + test/e2e/apimachinery/table_conversion.go:154 +[AfterEach] [sig-api-machinery] Servers with support for Table transformation test/e2e/framework/framework.go:187 -Dec 14 17:04:51.554: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-2193" for this suite. 12/14/22 17:04:51.561 -{"msg":"PASSED [sig-cli] Kubectl client Kubectl logs should be able to retrieve and filter logs [Conformance]","completed":315,"skipped":5926,"failed":0} +Feb 25 11:30:48.246: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "tables-2263" for this suite. 02/25/23 11:30:48.254 +{"msg":"PASSED [sig-api-machinery] Servers with support for Table transformation should return a 406 for a backend which does not implement metadata [Conformance]","completed":313,"skipped":6048,"failed":0} ------------------------------ -• [SLOW TEST] [7.054 seconds] -[sig-cli] Kubectl client -test/e2e/kubectl/framework.go:23 - Kubectl logs - test/e2e/kubectl/kubectl.go:1567 - should be able to retrieve and filter logs [Conformance] - test/e2e/kubectl/kubectl.go:1590 +• [0.063 seconds] +[sig-api-machinery] Servers with support for Table transformation +test/e2e/apimachinery/framework.go:23 + should return a 406 for a backend which does not implement metadata [Conformance] + test/e2e/apimachinery/table_conversion.go:154 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-cli] Kubectl client + [BeforeEach] [sig-api-machinery] Servers with support for Table transformation test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:04:44.52 - Dec 14 17:04:44.520: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 17:04:44.522 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:04:44.558 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:04:44.561 - [BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 - [BeforeEach] Kubectl logs - test/e2e/kubectl/kubectl.go:1570 - STEP: creating an pod 12/14/22 17:04:44.564 - Dec 14 17:04:44.567: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 run logs-generator --image=registry.k8s.io/e2e-test-images/agnhost:2.40 --restart=Never --pod-running-timeout=2m0s -- logs-generator --log-lines-total 100 --run-duration 20s' - Dec 14 17:04:44.736: INFO: stderr: "" - Dec 14 17:04:44.737: INFO: stdout: "pod/logs-generator created\n" - [It] should be able to retrieve and filter logs [Conformance] - test/e2e/kubectl/kubectl.go:1590 - STEP: Waiting for log generator to start. 12/14/22 17:04:44.737 - Dec 14 17:04:44.737: INFO: Waiting up to 5m0s for 1 pods to be running and ready, or succeeded: [logs-generator] - Dec 14 17:04:44.737: INFO: Waiting up to 5m0s for pod "logs-generator" in namespace "kubectl-2193" to be "running and ready, or succeeded" - Dec 14 17:04:44.747: INFO: Pod "logs-generator": Phase="Pending", Reason="", readiness=false. Elapsed: 9.956395ms - Dec 14 17:04:44.747: INFO: Error evaluating pod condition running and ready, or succeeded: want pod 'logs-generator' on 'kiedai4iakae-3' to be 'Running' but was 'Pending' - Dec 14 17:04:46.752: INFO: Pod "logs-generator": Phase="Running", Reason="", readiness=true. Elapsed: 2.014771871s - Dec 14 17:04:46.752: INFO: Pod "logs-generator" satisfied condition "running and ready, or succeeded" - Dec 14 17:04:46.752: INFO: Wanted all 1 pods to be running and ready, or succeeded. Result: true. Pods: [logs-generator] - STEP: checking for a matching strings 12/14/22 17:04:46.752 - Dec 14 17:04:46.752: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 logs logs-generator logs-generator' - Dec 14 17:04:46.919: INFO: stderr: "" - Dec 14 17:04:46.919: INFO: stdout: "I1214 17:04:45.628421 1 logs_generator.go:76] 0 GET /api/v1/namespaces/default/pods/lcq 517\nI1214 17:04:45.828378 1 logs_generator.go:76] 1 GET /api/v1/namespaces/default/pods/sp2 467\nI1214 17:04:46.028473 1 logs_generator.go:76] 2 POST /api/v1/namespaces/kube-system/pods/6jsk 469\nI1214 17:04:46.228919 1 logs_generator.go:76] 3 GET /api/v1/namespaces/default/pods/jrm 522\nI1214 17:04:46.429442 1 logs_generator.go:76] 4 POST /api/v1/namespaces/default/pods/c2w8 581\nI1214 17:04:46.629167 1 logs_generator.go:76] 5 POST /api/v1/namespaces/kube-system/pods/m6s 439\nI1214 17:04:46.829530 1 logs_generator.go:76] 6 GET /api/v1/namespaces/ns/pods/grxd 358\n" - STEP: limiting log lines 12/14/22 17:04:46.92 - Dec 14 17:04:46.920: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 logs logs-generator logs-generator --tail=1' - Dec 14 17:04:47.087: INFO: stderr: "" - Dec 14 17:04:47.087: INFO: stdout: "I1214 17:04:47.028916 1 logs_generator.go:76] 7 GET /api/v1/namespaces/kube-system/pods/p4ws 494\n" - Dec 14 17:04:47.087: INFO: got output "I1214 17:04:47.028916 1 logs_generator.go:76] 7 GET /api/v1/namespaces/kube-system/pods/p4ws 494\n" - STEP: limiting log bytes 12/14/22 17:04:47.087 - Dec 14 17:04:47.087: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 logs logs-generator logs-generator --limit-bytes=1' - Dec 14 17:04:47.366: INFO: stderr: "" - Dec 14 17:04:47.366: INFO: stdout: "I" - Dec 14 17:04:47.366: INFO: got output "I" - STEP: exposing timestamps 12/14/22 17:04:47.366 - Dec 14 17:04:47.367: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 logs logs-generator logs-generator --tail=1 --timestamps' - Dec 14 17:04:47.498: INFO: stderr: "" - Dec 14 17:04:47.498: INFO: stdout: "2022-12-14T17:04:47.428573564Z I1214 17:04:47.428503 1 logs_generator.go:76] 9 GET /api/v1/namespaces/ns/pods/6fl 376\n" - Dec 14 17:04:47.498: INFO: got output "2022-12-14T17:04:47.428573564Z I1214 17:04:47.428503 1 logs_generator.go:76] 9 GET /api/v1/namespaces/ns/pods/6fl 376\n" - STEP: restricting to a time range 12/14/22 17:04:47.498 - Dec 14 17:04:49.998: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 logs logs-generator logs-generator --since=1s' - Dec 14 17:04:50.239: INFO: stderr: "" - Dec 14 17:04:50.240: INFO: stdout: "I1214 17:04:49.429045 1 logs_generator.go:76] 19 POST /api/v1/namespaces/ns/pods/gqk 358\nI1214 17:04:49.628413 1 logs_generator.go:76] 20 GET /api/v1/namespaces/default/pods/h6f 478\nI1214 17:04:49.828797 1 logs_generator.go:76] 21 POST /api/v1/namespaces/default/pods/kwz 394\nI1214 17:04:50.029512 1 logs_generator.go:76] 22 POST /api/v1/namespaces/default/pods/6qlg 384\nI1214 17:04:50.228738 1 logs_generator.go:76] 23 PUT /api/v1/namespaces/default/pods/pnf 459\n" - Dec 14 17:04:50.240: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 logs logs-generator logs-generator --since=24h' - Dec 14 17:04:50.404: INFO: stderr: "" - Dec 14 17:04:50.404: INFO: stdout: "I1214 17:04:45.628421 1 logs_generator.go:76] 0 GET /api/v1/namespaces/default/pods/lcq 517\nI1214 17:04:45.828378 1 logs_generator.go:76] 1 GET /api/v1/namespaces/default/pods/sp2 467\nI1214 17:04:46.028473 1 logs_generator.go:76] 2 POST /api/v1/namespaces/kube-system/pods/6jsk 469\nI1214 17:04:46.228919 1 logs_generator.go:76] 3 GET /api/v1/namespaces/default/pods/jrm 522\nI1214 17:04:46.429442 1 logs_generator.go:76] 4 POST /api/v1/namespaces/default/pods/c2w8 581\nI1214 17:04:46.629167 1 logs_generator.go:76] 5 POST /api/v1/namespaces/kube-system/pods/m6s 439\nI1214 17:04:46.829530 1 logs_generator.go:76] 6 GET /api/v1/namespaces/ns/pods/grxd 358\nI1214 17:04:47.028916 1 logs_generator.go:76] 7 GET /api/v1/namespaces/kube-system/pods/p4ws 494\nI1214 17:04:47.228849 1 logs_generator.go:76] 8 POST /api/v1/namespaces/ns/pods/mhq 474\nI1214 17:04:47.428503 1 logs_generator.go:76] 9 GET /api/v1/namespaces/ns/pods/6fl 376\nI1214 17:04:47.629133 1 logs_generator.go:76] 10 GET /api/v1/namespaces/default/pods/lh6 454\nI1214 17:04:47.828368 1 logs_generator.go:76] 11 PUT /api/v1/namespaces/default/pods/k24n 382\nI1214 17:04:48.028922 1 logs_generator.go:76] 12 GET /api/v1/namespaces/ns/pods/7hzq 252\nI1214 17:04:48.228318 1 logs_generator.go:76] 13 GET /api/v1/namespaces/default/pods/n9jf 415\nI1214 17:04:48.429076 1 logs_generator.go:76] 14 GET /api/v1/namespaces/kube-system/pods/9df 378\nI1214 17:04:48.628403 1 logs_generator.go:76] 15 PUT /api/v1/namespaces/default/pods/sj7w 521\nI1214 17:04:48.828790 1 logs_generator.go:76] 16 GET /api/v1/namespaces/default/pods/vvnl 542\nI1214 17:04:49.029166 1 logs_generator.go:76] 17 PUT /api/v1/namespaces/ns/pods/hdmn 541\nI1214 17:04:49.228478 1 logs_generator.go:76] 18 PUT /api/v1/namespaces/kube-system/pods/ct6 414\nI1214 17:04:49.429045 1 logs_generator.go:76] 19 POST /api/v1/namespaces/ns/pods/gqk 358\nI1214 17:04:49.628413 1 logs_generator.go:76] 20 GET /api/v1/namespaces/default/pods/h6f 478\nI1214 17:04:49.828797 1 logs_generator.go:76] 21 POST /api/v1/namespaces/default/pods/kwz 394\nI1214 17:04:50.029512 1 logs_generator.go:76] 22 POST /api/v1/namespaces/default/pods/6qlg 384\nI1214 17:04:50.228738 1 logs_generator.go:76] 23 PUT /api/v1/namespaces/default/pods/pnf 459\n" - [AfterEach] Kubectl logs - test/e2e/kubectl/kubectl.go:1575 - Dec 14 17:04:50.405: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-2193 delete pod logs-generator' - Dec 14 17:04:51.553: INFO: stderr: "" - Dec 14 17:04:51.553: INFO: stdout: "pod \"logs-generator\" deleted\n" - [AfterEach] [sig-cli] Kubectl client + STEP: Creating a kubernetes client 02/25/23 11:30:48.203 + Feb 25 11:30:48.203: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename tables 02/25/23 11:30:48.207 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:30:48.234 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:30:48.238 + [BeforeEach] [sig-api-machinery] Servers with support for Table transformation + test/e2e/apimachinery/table_conversion.go:49 + [It] should return a 406 for a backend which does not implement metadata [Conformance] + test/e2e/apimachinery/table_conversion.go:154 + [AfterEach] [sig-api-machinery] Servers with support for Table transformation test/e2e/framework/framework.go:187 - Dec 14 17:04:51.554: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-2193" for this suite. 12/14/22 17:04:51.561 + Feb 25 11:30:48.246: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "tables-2263" for this suite. 02/25/23 11:30:48.254 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSS +SSSSS ------------------------------ -[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] - Should recreate evicted statefulset [Conformance] - test/e2e/apps/statefulset.go:737 -[BeforeEach] [sig-apps] StatefulSet +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + should unconditionally reject operations on fail closed webhook [Conformance] + test/e2e/apimachinery/webhook.go:238 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:04:51.582 -Dec 14 17:04:51.583: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename statefulset 12/14/22 17:04:51.586 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:04:51.615 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:04:51.619 -[BeforeEach] [sig-apps] StatefulSet - test/e2e/apps/statefulset.go:96 -[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:111 -STEP: Creating service test in namespace statefulset-159 12/14/22 17:04:51.622 -[It] Should recreate evicted statefulset [Conformance] - test/e2e/apps/statefulset.go:737 -STEP: Looking for a node to schedule stateful set and pod 12/14/22 17:04:51.63 -STEP: Creating pod with conflicting port in namespace statefulset-159 12/14/22 17:04:51.637 -STEP: Waiting until pod test-pod will start running in namespace statefulset-159 12/14/22 17:04:51.65 -Dec 14 17:04:51.650: INFO: Waiting up to 5m0s for pod "test-pod" in namespace "statefulset-159" to be "running" -Dec 14 17:04:51.654: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 4.029846ms -Dec 14 17:04:53.662: INFO: Pod "test-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.012193953s -Dec 14 17:04:53.662: INFO: Pod "test-pod" satisfied condition "running" -STEP: Creating statefulset with conflicting port in namespace statefulset-159 12/14/22 17:04:53.662 -STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-159 12/14/22 17:04:53.675 -Dec 14 17:04:53.722: INFO: Observed stateful pod in namespace: statefulset-159, name: ss-0, uid: 659c2af9-9408-4f24-a171-38f80eb27ee3, status phase: Pending. Waiting for statefulset controller to delete. -Dec 14 17:04:53.752: INFO: Observed stateful pod in namespace: statefulset-159, name: ss-0, uid: 659c2af9-9408-4f24-a171-38f80eb27ee3, status phase: Failed. Waiting for statefulset controller to delete. -Dec 14 17:04:53.803: INFO: Observed stateful pod in namespace: statefulset-159, name: ss-0, uid: 659c2af9-9408-4f24-a171-38f80eb27ee3, status phase: Failed. Waiting for statefulset controller to delete. -Dec 14 17:04:53.809: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-159 -STEP: Removing pod with conflicting port in namespace statefulset-159 12/14/22 17:04:53.809 -STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-159 and will be in running state 12/14/22 17:04:53.839 -[AfterEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:122 -Dec 14 17:04:57.885: INFO: Deleting all statefulset in ns statefulset-159 -Dec 14 17:04:57.889: INFO: Scaling statefulset ss to 0 -Dec 14 17:05:07.914: INFO: Waiting for statefulset status.replicas updated to 0 -Dec 14 17:05:07.918: INFO: Deleting statefulset ss -[AfterEach] [sig-apps] StatefulSet +STEP: Creating a kubernetes client 02/25/23 11:30:48.267 +Feb 25 11:30:48.267: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 11:30:48.269 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:30:48.3 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:30:48.303 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 +STEP: Setting up server cert 02/25/23 11:30:48.334 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 11:30:49.49 +STEP: Deploying the webhook pod 02/25/23 11:30:49.503 +STEP: Wait for the deployment to be ready 02/25/23 11:30:49.523 +Feb 25 11:30:49.540: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set +Feb 25 11:30:51.561: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 11, 30, 49, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 30, 49, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 11, 30, 49, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 30, 49, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} +STEP: Deploying the webhook service 02/25/23 11:30:53.567 +STEP: Verifying the service has paired with the endpoint 02/25/23 11:30:53.589 +Feb 25 11:30:54.589: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should unconditionally reject operations on fail closed webhook [Conformance] + test/e2e/apimachinery/webhook.go:238 +STEP: Registering a webhook that server cannot talk to, with fail closed policy, via the AdmissionRegistration API 02/25/23 11:30:54.597 +STEP: create a namespace for the webhook 02/25/23 11:30:54.628 +STEP: create a configmap should be unconditionally rejected by the webhook 02/25/23 11:30:54.645 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 17:05:07.935: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "statefulset-159" for this suite. 12/14/22 17:05:07.94 -{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] Should recreate evicted statefulset [Conformance]","completed":316,"skipped":5943,"failed":0} +Feb 25 11:30:54.700: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-2849" for this suite. 02/25/23 11:30:54.707 +STEP: Destroying namespace "webhook-2849-markers" for this suite. 02/25/23 11:30:54.717 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should unconditionally reject operations on fail closed webhook [Conformance]","completed":314,"skipped":6053,"failed":0} ------------------------------ -• [SLOW TEST] [16.366 seconds] -[sig-apps] StatefulSet -test/e2e/apps/framework.go:23 - Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:101 - Should recreate evicted statefulset [Conformance] - test/e2e/apps/statefulset.go:737 +• [SLOW TEST] [6.555 seconds] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + should unconditionally reject operations on fail closed webhook [Conformance] + test/e2e/apimachinery/webhook.go:238 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] StatefulSet + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:04:51.582 - Dec 14 17:04:51.583: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename statefulset 12/14/22 17:04:51.586 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:04:51.615 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:04:51.619 - [BeforeEach] [sig-apps] StatefulSet - test/e2e/apps/statefulset.go:96 - [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:111 - STEP: Creating service test in namespace statefulset-159 12/14/22 17:04:51.622 - [It] Should recreate evicted statefulset [Conformance] - test/e2e/apps/statefulset.go:737 - STEP: Looking for a node to schedule stateful set and pod 12/14/22 17:04:51.63 - STEP: Creating pod with conflicting port in namespace statefulset-159 12/14/22 17:04:51.637 - STEP: Waiting until pod test-pod will start running in namespace statefulset-159 12/14/22 17:04:51.65 - Dec 14 17:04:51.650: INFO: Waiting up to 5m0s for pod "test-pod" in namespace "statefulset-159" to be "running" - Dec 14 17:04:51.654: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 4.029846ms - Dec 14 17:04:53.662: INFO: Pod "test-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.012193953s - Dec 14 17:04:53.662: INFO: Pod "test-pod" satisfied condition "running" - STEP: Creating statefulset with conflicting port in namespace statefulset-159 12/14/22 17:04:53.662 - STEP: Waiting until stateful pod ss-0 will be recreated and deleted at least once in namespace statefulset-159 12/14/22 17:04:53.675 - Dec 14 17:04:53.722: INFO: Observed stateful pod in namespace: statefulset-159, name: ss-0, uid: 659c2af9-9408-4f24-a171-38f80eb27ee3, status phase: Pending. Waiting for statefulset controller to delete. - Dec 14 17:04:53.752: INFO: Observed stateful pod in namespace: statefulset-159, name: ss-0, uid: 659c2af9-9408-4f24-a171-38f80eb27ee3, status phase: Failed. Waiting for statefulset controller to delete. - Dec 14 17:04:53.803: INFO: Observed stateful pod in namespace: statefulset-159, name: ss-0, uid: 659c2af9-9408-4f24-a171-38f80eb27ee3, status phase: Failed. Waiting for statefulset controller to delete. - Dec 14 17:04:53.809: INFO: Observed delete event for stateful pod ss-0 in namespace statefulset-159 - STEP: Removing pod with conflicting port in namespace statefulset-159 12/14/22 17:04:53.809 - STEP: Waiting when stateful pod ss-0 will be recreated in namespace statefulset-159 and will be in running state 12/14/22 17:04:53.839 - [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] - test/e2e/apps/statefulset.go:122 - Dec 14 17:04:57.885: INFO: Deleting all statefulset in ns statefulset-159 - Dec 14 17:04:57.889: INFO: Scaling statefulset ss to 0 - Dec 14 17:05:07.914: INFO: Waiting for statefulset status.replicas updated to 0 - Dec 14 17:05:07.918: INFO: Deleting statefulset ss - [AfterEach] [sig-apps] StatefulSet + STEP: Creating a kubernetes client 02/25/23 11:30:48.267 + Feb 25 11:30:48.267: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 11:30:48.269 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:30:48.3 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:30:48.303 + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 + STEP: Setting up server cert 02/25/23 11:30:48.334 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 11:30:49.49 + STEP: Deploying the webhook pod 02/25/23 11:30:49.503 + STEP: Wait for the deployment to be ready 02/25/23 11:30:49.523 + Feb 25 11:30:49.540: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set + Feb 25 11:30:51.561: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 11, 30, 49, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 30, 49, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 11, 30, 49, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 30, 49, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} + STEP: Deploying the webhook service 02/25/23 11:30:53.567 + STEP: Verifying the service has paired with the endpoint 02/25/23 11:30:53.589 + Feb 25 11:30:54.589: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] should unconditionally reject operations on fail closed webhook [Conformance] + test/e2e/apimachinery/webhook.go:238 + STEP: Registering a webhook that server cannot talk to, with fail closed policy, via the AdmissionRegistration API 02/25/23 11:30:54.597 + STEP: create a namespace for the webhook 02/25/23 11:30:54.628 + STEP: create a configmap should be unconditionally rejected by the webhook 02/25/23 11:30:54.645 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 17:05:07.935: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "statefulset-159" for this suite. 12/14/22 17:05:07.94 + Feb 25 11:30:54.700: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-2849" for this suite. 02/25/23 11:30:54.707 + STEP: Destroying namespace "webhook-2849-markers" for this suite. 02/25/23 11:30:54.717 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSS +SSSSSSSSSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:146 -[BeforeEach] [sig-storage] EmptyDir volumes +[sig-node] Security Context when creating containers with AllowPrivilegeEscalation + should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:608 +[BeforeEach] [sig-node] Security Context test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:05:07.95 -Dec 14 17:05:07.950: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir 12/14/22 17:05:07.951 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:07.977 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:07.983 -[It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:146 -STEP: Creating a pod to test emptydir 0777 on tmpfs 12/14/22 17:05:07.987 -Dec 14 17:05:08.001: INFO: Waiting up to 5m0s for pod "pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d" in namespace "emptydir-919" to be "Succeeded or Failed" -Dec 14 17:05:08.009: INFO: Pod "pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d": Phase="Pending", Reason="", readiness=false. Elapsed: 7.971884ms -Dec 14 17:05:10.015: INFO: Pod "pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013370124s -Dec 14 17:05:12.015: INFO: Pod "pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013380526s -STEP: Saw pod success 12/14/22 17:05:12.015 -Dec 14 17:05:12.015: INFO: Pod "pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d" satisfied condition "Succeeded or Failed" -Dec 14 17:05:12.020: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d container test-container: -STEP: delete the pod 12/14/22 17:05:12.247 -Dec 14 17:05:12.266: INFO: Waiting for pod pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d to disappear -Dec 14 17:05:12.270: INFO: Pod pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +STEP: Creating a kubernetes client 02/25/23 11:30:54.825 +Feb 25 11:30:54.825: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename security-context-test 02/25/23 11:30:54.831 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:30:54.887 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:30:54.892 +[BeforeEach] [sig-node] Security Context + test/e2e/common/node/security_context.go:49 +[It] should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:608 +Feb 25 11:31:04.678: INFO: Waiting up to 5m0s for pod "alpine-nnp-false-65706326-d54c-47bc-958f-726b4df842a2" in namespace "security-context-test-1960" to be "Succeeded or Failed" +Feb 25 11:31:04.684: INFO: Pod "alpine-nnp-false-65706326-d54c-47bc-958f-726b4df842a2": Phase="Pending", Reason="", readiness=false. Elapsed: 5.069011ms +Feb 25 11:31:06.695: INFO: Pod "alpine-nnp-false-65706326-d54c-47bc-958f-726b4df842a2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01636305s +Feb 25 11:31:08.691: INFO: Pod "alpine-nnp-false-65706326-d54c-47bc-958f-726b4df842a2": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012082461s +Feb 25 11:31:10.691: INFO: Pod "alpine-nnp-false-65706326-d54c-47bc-958f-726b4df842a2": Phase="Pending", Reason="", readiness=false. Elapsed: 6.012168335s +Feb 25 11:31:12.690: INFO: Pod "alpine-nnp-false-65706326-d54c-47bc-958f-726b4df842a2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.011643171s +Feb 25 11:31:12.690: INFO: Pod "alpine-nnp-false-65706326-d54c-47bc-958f-726b4df842a2" satisfied condition "Succeeded or Failed" +[AfterEach] [sig-node] Security Context test/e2e/framework/framework.go:187 -Dec 14 17:05:12.270: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-919" for this suite. 12/14/22 17:05:12.275 -{"msg":"PASSED [sig-storage] EmptyDir volumes should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance]","completed":317,"skipped":5950,"failed":0} +Feb 25 11:31:12.710: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "security-context-test-1960" for this suite. 02/25/23 11:31:12.717 +{"msg":"PASSED [sig-node] Security Context when creating containers with AllowPrivilegeEscalation should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance]","completed":315,"skipped":6067,"failed":0} ------------------------------ -• [4.332 seconds] -[sig-storage] EmptyDir volumes -test/e2e/common/storage/framework.go:23 - should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:146 +• [SLOW TEST] [17.904 seconds] +[sig-node] Security Context +test/e2e/common/node/framework.go:23 + when creating containers with AllowPrivilegeEscalation + test/e2e/common/node/security_context.go:554 + should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:608 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-node] Security Context + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 11:30:54.825 + Feb 25 11:30:54.825: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename security-context-test 02/25/23 11:30:54.831 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:30:54.887 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:30:54.892 + [BeforeEach] [sig-node] Security Context + test/e2e/common/node/security_context.go:49 + [It] should not allow privilege escalation when false [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:608 + Feb 25 11:31:04.678: INFO: Waiting up to 5m0s for pod "alpine-nnp-false-65706326-d54c-47bc-958f-726b4df842a2" in namespace "security-context-test-1960" to be "Succeeded or Failed" + Feb 25 11:31:04.684: INFO: Pod "alpine-nnp-false-65706326-d54c-47bc-958f-726b4df842a2": Phase="Pending", Reason="", readiness=false. Elapsed: 5.069011ms + Feb 25 11:31:06.695: INFO: Pod "alpine-nnp-false-65706326-d54c-47bc-958f-726b4df842a2": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01636305s + Feb 25 11:31:08.691: INFO: Pod "alpine-nnp-false-65706326-d54c-47bc-958f-726b4df842a2": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012082461s + Feb 25 11:31:10.691: INFO: Pod "alpine-nnp-false-65706326-d54c-47bc-958f-726b4df842a2": Phase="Pending", Reason="", readiness=false. Elapsed: 6.012168335s + Feb 25 11:31:12.690: INFO: Pod "alpine-nnp-false-65706326-d54c-47bc-958f-726b4df842a2": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.011643171s + Feb 25 11:31:12.690: INFO: Pod "alpine-nnp-false-65706326-d54c-47bc-958f-726b4df842a2" satisfied condition "Succeeded or Failed" + [AfterEach] [sig-node] Security Context + test/e2e/framework/framework.go:187 + Feb 25 11:31:12.710: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "security-context-test-1960" for this suite. 02/25/23 11:31:12.717 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSSSSS +------------------------------ +[sig-api-machinery] Namespaces [Serial] + should ensure that all pods are removed when a namespace is deleted [Conformance] + test/e2e/apimachinery/namespace.go:242 +[BeforeEach] [sig-api-machinery] Namespaces [Serial] + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 11:31:12.737 +Feb 25 11:31:12.737: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename namespaces 02/25/23 11:31:12.743 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:31:12.767 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:31:12.771 +[It] should ensure that all pods are removed when a namespace is deleted [Conformance] + test/e2e/apimachinery/namespace.go:242 +STEP: Creating a test namespace 02/25/23 11:31:12.776 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:31:12.812 +STEP: Creating a pod in the namespace 02/25/23 11:31:12.815 +STEP: Waiting for the pod to have running status 02/25/23 11:31:12.827 +Feb 25 11:31:12.827: INFO: Waiting up to 5m0s for pod "test-pod" in namespace "nsdeletetest-9580" to be "running" +Feb 25 11:31:12.838: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 10.311042ms +Feb 25 11:31:14.846: INFO: Pod "test-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.019103993s +Feb 25 11:31:14.846: INFO: Pod "test-pod" satisfied condition "running" +STEP: Deleting the namespace 02/25/23 11:31:14.846 +STEP: Waiting for the namespace to be removed. 02/25/23 11:31:14.859 +STEP: Recreating the namespace 02/25/23 11:31:25.865 +STEP: Verifying there are no pods in the namespace 02/25/23 11:31:25.887 +[AfterEach] [sig-api-machinery] Namespaces [Serial] + test/e2e/framework/framework.go:187 +Feb 25 11:31:25.892: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "namespaces-1970" for this suite. 02/25/23 11:31:25.899 +STEP: Destroying namespace "nsdeletetest-9580" for this suite. 02/25/23 11:31:25.909 +Feb 25 11:31:25.913: INFO: Namespace nsdeletetest-9580 was already deleted +STEP: Destroying namespace "nsdeletetest-5776" for this suite. 02/25/23 11:31:25.913 +{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all pods are removed when a namespace is deleted [Conformance]","completed":316,"skipped":6085,"failed":0} +------------------------------ +• [SLOW TEST] [13.186 seconds] +[sig-api-machinery] Namespaces [Serial] +test/e2e/apimachinery/framework.go:23 + should ensure that all pods are removed when a namespace is deleted [Conformance] + test/e2e/apimachinery/namespace.go:242 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir volumes + [BeforeEach] [sig-api-machinery] Namespaces [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:05:07.95 - Dec 14 17:05:07.950: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir 12/14/22 17:05:07.951 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:07.977 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:07.983 - [It] should support (non-root,0777,tmpfs) [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:146 - STEP: Creating a pod to test emptydir 0777 on tmpfs 12/14/22 17:05:07.987 - Dec 14 17:05:08.001: INFO: Waiting up to 5m0s for pod "pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d" in namespace "emptydir-919" to be "Succeeded or Failed" - Dec 14 17:05:08.009: INFO: Pod "pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d": Phase="Pending", Reason="", readiness=false. Elapsed: 7.971884ms - Dec 14 17:05:10.015: INFO: Pod "pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013370124s - Dec 14 17:05:12.015: INFO: Pod "pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013380526s - STEP: Saw pod success 12/14/22 17:05:12.015 - Dec 14 17:05:12.015: INFO: Pod "pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d" satisfied condition "Succeeded or Failed" - Dec 14 17:05:12.020: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d container test-container: - STEP: delete the pod 12/14/22 17:05:12.247 - Dec 14 17:05:12.266: INFO: Waiting for pod pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d to disappear - Dec 14 17:05:12.270: INFO: Pod pod-316b6cf0-9908-4a7d-933b-e7aa4a54915d no longer exists - [AfterEach] [sig-storage] EmptyDir volumes + STEP: Creating a kubernetes client 02/25/23 11:31:12.737 + Feb 25 11:31:12.737: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename namespaces 02/25/23 11:31:12.743 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:31:12.767 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:31:12.771 + [It] should ensure that all pods are removed when a namespace is deleted [Conformance] + test/e2e/apimachinery/namespace.go:242 + STEP: Creating a test namespace 02/25/23 11:31:12.776 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:31:12.812 + STEP: Creating a pod in the namespace 02/25/23 11:31:12.815 + STEP: Waiting for the pod to have running status 02/25/23 11:31:12.827 + Feb 25 11:31:12.827: INFO: Waiting up to 5m0s for pod "test-pod" in namespace "nsdeletetest-9580" to be "running" + Feb 25 11:31:12.838: INFO: Pod "test-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 10.311042ms + Feb 25 11:31:14.846: INFO: Pod "test-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.019103993s + Feb 25 11:31:14.846: INFO: Pod "test-pod" satisfied condition "running" + STEP: Deleting the namespace 02/25/23 11:31:14.846 + STEP: Waiting for the namespace to be removed. 02/25/23 11:31:14.859 + STEP: Recreating the namespace 02/25/23 11:31:25.865 + STEP: Verifying there are no pods in the namespace 02/25/23 11:31:25.887 + [AfterEach] [sig-api-machinery] Namespaces [Serial] test/e2e/framework/framework.go:187 - Dec 14 17:05:12.270: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-919" for this suite. 12/14/22 17:05:12.275 + Feb 25 11:31:25.892: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "namespaces-1970" for this suite. 02/25/23 11:31:25.899 + STEP: Destroying namespace "nsdeletetest-9580" for this suite. 02/25/23 11:31:25.909 + Feb 25 11:31:25.913: INFO: Namespace nsdeletetest-9580 was already deleted + STEP: Destroying namespace "nsdeletetest-5776" for this suite. 02/25/23 11:31:25.913 << End Captured GinkgoWriter Output ------------------------------ -SSSS +SSSSSSSSS ------------------------------ -[sig-cli] Kubectl client Kubectl label - should update the label on a resource [Conformance] - test/e2e/kubectl/kubectl.go:1507 -[BeforeEach] [sig-cli] Kubectl client +[sig-node] Lease + lease API should be available [Conformance] + test/e2e/common/node/lease.go:72 +[BeforeEach] [sig-node] Lease test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:05:12.282 -Dec 14 17:05:12.283: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 17:05:12.284 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:12.304 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:12.306 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 -[BeforeEach] Kubectl label - test/e2e/kubectl/kubectl.go:1492 -STEP: creating the pod 12/14/22 17:05:12.309 -Dec 14 17:05:12.310: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 create -f -' -Dec 14 17:05:13.537: INFO: stderr: "" -Dec 14 17:05:13.538: INFO: stdout: "pod/pause created\n" -Dec 14 17:05:13.538: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause] -Dec 14 17:05:13.538: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-4358" to be "running and ready" -Dec 14 17:05:13.542: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 3.98248ms -Dec 14 17:05:13.542: INFO: Error evaluating pod condition running and ready: want pod 'pause' on '' to be 'Running' but was 'Pending' -Dec 14 17:05:15.548: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 2.01078533s -Dec 14 17:05:15.548: INFO: Pod "pause" satisfied condition "running and ready" -Dec 14 17:05:15.548: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause] -[It] should update the label on a resource [Conformance] - test/e2e/kubectl/kubectl.go:1507 -STEP: adding the label testing-label with value testing-label-value to a pod 12/14/22 17:05:15.549 -Dec 14 17:05:15.549: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 label pods pause testing-label=testing-label-value' -Dec 14 17:05:15.669: INFO: stderr: "" -Dec 14 17:05:15.669: INFO: stdout: "pod/pause labeled\n" -STEP: verifying the pod has the label testing-label with the value testing-label-value 12/14/22 17:05:15.67 -Dec 14 17:05:15.670: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 get pod pause -L testing-label' -Dec 14 17:05:15.817: INFO: stderr: "" -Dec 14 17:05:15.817: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 2s testing-label-value\n" -STEP: removing the label testing-label of a pod 12/14/22 17:05:15.817 -Dec 14 17:05:15.817: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 label pods pause testing-label-' -Dec 14 17:05:15.965: INFO: stderr: "" -Dec 14 17:05:15.965: INFO: stdout: "pod/pause unlabeled\n" -STEP: verifying the pod doesn't have the label testing-label 12/14/22 17:05:15.965 -Dec 14 17:05:15.966: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 get pod pause -L testing-label' -Dec 14 17:05:16.116: INFO: stderr: "" -Dec 14 17:05:16.116: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 3s \n" -[AfterEach] Kubectl label - test/e2e/kubectl/kubectl.go:1498 -STEP: using delete to clean up resources 12/14/22 17:05:16.116 -Dec 14 17:05:16.117: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 delete --grace-period=0 --force -f -' -Dec 14 17:05:16.289: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Dec 14 17:05:16.289: INFO: stdout: "pod \"pause\" force deleted\n" -Dec 14 17:05:16.289: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 get rc,svc -l name=pause --no-headers' -Dec 14 17:05:16.506: INFO: stderr: "No resources found in kubectl-4358 namespace.\n" -Dec 14 17:05:16.506: INFO: stdout: "" -Dec 14 17:05:16.507: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 get pods -l name=pause -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' -Dec 14 17:05:16.708: INFO: stderr: "" -Dec 14 17:05:16.708: INFO: stdout: "" -[AfterEach] [sig-cli] Kubectl client +STEP: Creating a kubernetes client 02/25/23 11:31:25.925 +Feb 25 11:31:25.926: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename lease-test 02/25/23 11:31:25.928 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:31:25.953 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:31:25.957 +[It] lease API should be available [Conformance] + test/e2e/common/node/lease.go:72 +[AfterEach] [sig-node] Lease test/e2e/framework/framework.go:187 -Dec 14 17:05:16.708: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-4358" for this suite. 12/14/22 17:05:16.716 -{"msg":"PASSED [sig-cli] Kubectl client Kubectl label should update the label on a resource [Conformance]","completed":318,"skipped":5954,"failed":0} +Feb 25 11:31:26.040: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "lease-test-9992" for this suite. 02/25/23 11:31:26.047 +{"msg":"PASSED [sig-node] Lease lease API should be available [Conformance]","completed":317,"skipped":6094,"failed":0} ------------------------------ -• [4.442 seconds] -[sig-cli] Kubectl client -test/e2e/kubectl/framework.go:23 - Kubectl label - test/e2e/kubectl/kubectl.go:1490 - should update the label on a resource [Conformance] - test/e2e/kubectl/kubectl.go:1507 +• [0.130 seconds] +[sig-node] Lease +test/e2e/common/node/framework.go:23 + lease API should be available [Conformance] + test/e2e/common/node/lease.go:72 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-cli] Kubectl client + [BeforeEach] [sig-node] Lease test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:05:12.282 - Dec 14 17:05:12.283: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 17:05:12.284 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:12.304 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:12.306 - [BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 - [BeforeEach] Kubectl label - test/e2e/kubectl/kubectl.go:1492 - STEP: creating the pod 12/14/22 17:05:12.309 - Dec 14 17:05:12.310: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 create -f -' - Dec 14 17:05:13.537: INFO: stderr: "" - Dec 14 17:05:13.538: INFO: stdout: "pod/pause created\n" - Dec 14 17:05:13.538: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause] - Dec 14 17:05:13.538: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-4358" to be "running and ready" - Dec 14 17:05:13.542: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 3.98248ms - Dec 14 17:05:13.542: INFO: Error evaluating pod condition running and ready: want pod 'pause' on '' to be 'Running' but was 'Pending' - Dec 14 17:05:15.548: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 2.01078533s - Dec 14 17:05:15.548: INFO: Pod "pause" satisfied condition "running and ready" - Dec 14 17:05:15.548: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause] - [It] should update the label on a resource [Conformance] - test/e2e/kubectl/kubectl.go:1507 - STEP: adding the label testing-label with value testing-label-value to a pod 12/14/22 17:05:15.549 - Dec 14 17:05:15.549: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 label pods pause testing-label=testing-label-value' - Dec 14 17:05:15.669: INFO: stderr: "" - Dec 14 17:05:15.669: INFO: stdout: "pod/pause labeled\n" - STEP: verifying the pod has the label testing-label with the value testing-label-value 12/14/22 17:05:15.67 - Dec 14 17:05:15.670: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 get pod pause -L testing-label' - Dec 14 17:05:15.817: INFO: stderr: "" - Dec 14 17:05:15.817: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 2s testing-label-value\n" - STEP: removing the label testing-label of a pod 12/14/22 17:05:15.817 - Dec 14 17:05:15.817: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 label pods pause testing-label-' - Dec 14 17:05:15.965: INFO: stderr: "" - Dec 14 17:05:15.965: INFO: stdout: "pod/pause unlabeled\n" - STEP: verifying the pod doesn't have the label testing-label 12/14/22 17:05:15.965 - Dec 14 17:05:15.966: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 get pod pause -L testing-label' - Dec 14 17:05:16.116: INFO: stderr: "" - Dec 14 17:05:16.116: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 3s \n" - [AfterEach] Kubectl label - test/e2e/kubectl/kubectl.go:1498 - STEP: using delete to clean up resources 12/14/22 17:05:16.116 - Dec 14 17:05:16.117: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 delete --grace-period=0 --force -f -' - Dec 14 17:05:16.289: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" - Dec 14 17:05:16.289: INFO: stdout: "pod \"pause\" force deleted\n" - Dec 14 17:05:16.289: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 get rc,svc -l name=pause --no-headers' - Dec 14 17:05:16.506: INFO: stderr: "No resources found in kubectl-4358 namespace.\n" - Dec 14 17:05:16.506: INFO: stdout: "" - Dec 14 17:05:16.507: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-4358 get pods -l name=pause -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' - Dec 14 17:05:16.708: INFO: stderr: "" - Dec 14 17:05:16.708: INFO: stdout: "" - [AfterEach] [sig-cli] Kubectl client + STEP: Creating a kubernetes client 02/25/23 11:31:25.925 + Feb 25 11:31:25.926: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename lease-test 02/25/23 11:31:25.928 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:31:25.953 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:31:25.957 + [It] lease API should be available [Conformance] + test/e2e/common/node/lease.go:72 + [AfterEach] [sig-node] Lease test/e2e/framework/framework.go:187 - Dec 14 17:05:16.708: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-4358" for this suite. 12/14/22 17:05:16.716 + Feb 25 11:31:26.040: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "lease-test-9992" for this suite. 02/25/23 11:31:26.047 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSS +SSSS ------------------------------ -[sig-apps] Daemon set [Serial] - should retry creating failed daemon pods [Conformance] - test/e2e/apps/daemon_set.go:293 -[BeforeEach] [sig-apps] Daemon set [Serial] +[sig-node] Probing container + should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:148 +[BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:05:16.728 -Dec 14 17:05:16.729: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename daemonsets 12/14/22 17:05:16.73 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:16.75 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:16.753 -[BeforeEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:145 -[It] should retry creating failed daemon pods [Conformance] - test/e2e/apps/daemon_set.go:293 -STEP: Creating a simple DaemonSet "daemon-set" 12/14/22 17:05:16.779 -STEP: Check that daemon pods launch on every node of the cluster. 12/14/22 17:05:16.793 -Dec 14 17:05:16.802: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 17:05:16.802: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 17:05:17.842: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 -Dec 14 17:05:17.851: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 17:05:18.816: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 -Dec 14 17:05:18.816: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 -Dec 14 17:05:19.824: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 -Dec 14 17:05:19.824: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set -STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived. 12/14/22 17:05:19.832 -Dec 14 17:05:19.880: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 -Dec 14 17:05:19.880: INFO: Node kiedai4iakae-3 is running 0 daemon pod, expected 1 -Dec 14 17:05:20.891: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 -Dec 14 17:05:20.891: INFO: Node kiedai4iakae-3 is running 0 daemon pod, expected 1 -Dec 14 17:05:21.896: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 -Dec 14 17:05:21.896: INFO: Node kiedai4iakae-3 is running 0 daemon pod, expected 1 -Dec 14 17:05:22.896: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 -Dec 14 17:05:22.896: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set -STEP: Wait for the failed daemon pod to be completely deleted. 12/14/22 17:05:22.896 -[AfterEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:110 -STEP: Deleting DaemonSet "daemon-set" 12/14/22 17:05:22.906 -STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-7350, will wait for the garbage collector to delete the pods 12/14/22 17:05:22.906 -Dec 14 17:05:22.970: INFO: Deleting DaemonSet.extensions daemon-set took: 8.007462ms -Dec 14 17:05:23.071: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.999927ms -Dec 14 17:05:25.782: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 -Dec 14 17:05:25.782: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set -Dec 14 17:05:25.789: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"29734"},"items":null} - -Dec 14 17:05:25.800: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"29734"},"items":null} - -[AfterEach] [sig-apps] Daemon set [Serial] +STEP: Creating a kubernetes client 02/25/23 11:31:26.056 +Feb 25 11:31:26.057: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-probe 02/25/23 11:31:26.059 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:31:26.082 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:31:26.086 +[BeforeEach] [sig-node] Probing container + test/e2e/common/node/container_probe.go:59 +[It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:148 +STEP: Creating pod busybox-de151162-692a-4290-832a-6281a820b040 in namespace container-probe-5837 02/25/23 11:31:26.09 +Feb 25 11:31:26.100: INFO: Waiting up to 5m0s for pod "busybox-de151162-692a-4290-832a-6281a820b040" in namespace "container-probe-5837" to be "not pending" +Feb 25 11:31:26.107: INFO: Pod "busybox-de151162-692a-4290-832a-6281a820b040": Phase="Pending", Reason="", readiness=false. Elapsed: 6.610336ms +Feb 25 11:31:28.112: INFO: Pod "busybox-de151162-692a-4290-832a-6281a820b040": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011684075s +Feb 25 11:31:30.113: INFO: Pod "busybox-de151162-692a-4290-832a-6281a820b040": Phase="Running", Reason="", readiness=true. Elapsed: 4.012813538s +Feb 25 11:31:30.113: INFO: Pod "busybox-de151162-692a-4290-832a-6281a820b040" satisfied condition "not pending" +Feb 25 11:31:30.113: INFO: Started pod busybox-de151162-692a-4290-832a-6281a820b040 in namespace container-probe-5837 +STEP: checking the pod's current state and verifying that restartCount is present 02/25/23 11:31:30.113 +Feb 25 11:31:30.118: INFO: Initial restart count of pod busybox-de151162-692a-4290-832a-6281a820b040 is 0 +STEP: deleting the pod 02/25/23 11:35:31.25 +[AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 -Dec 14 17:05:25.819: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "daemonsets-7350" for this suite. 12/14/22 17:05:25.825 -{"msg":"PASSED [sig-apps] Daemon set [Serial] should retry creating failed daemon pods [Conformance]","completed":319,"skipped":5980,"failed":0} +Feb 25 11:35:31.285: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-probe-5837" for this suite. 02/25/23 11:35:31.296 +{"msg":"PASSED [sig-node] Probing container should *not* be restarted with a exec \"cat /tmp/health\" liveness probe [NodeConformance] [Conformance]","completed":318,"skipped":6098,"failed":0} ------------------------------ -• [SLOW TEST] [9.107 seconds] -[sig-apps] Daemon set [Serial] -test/e2e/apps/framework.go:23 - should retry creating failed daemon pods [Conformance] - test/e2e/apps/daemon_set.go:293 +• [SLOW TEST] [245.261 seconds] +[sig-node] Probing container +test/e2e/common/node/framework.go:23 + should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:148 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Daemon set [Serial] + [BeforeEach] [sig-node] Probing container test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:05:16.728 - Dec 14 17:05:16.729: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename daemonsets 12/14/22 17:05:16.73 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:16.75 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:16.753 - [BeforeEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:145 - [It] should retry creating failed daemon pods [Conformance] - test/e2e/apps/daemon_set.go:293 - STEP: Creating a simple DaemonSet "daemon-set" 12/14/22 17:05:16.779 - STEP: Check that daemon pods launch on every node of the cluster. 12/14/22 17:05:16.793 - Dec 14 17:05:16.802: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 17:05:16.802: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 17:05:17.842: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 1 - Dec 14 17:05:17.851: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 17:05:18.816: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 - Dec 14 17:05:18.816: INFO: Node kiedai4iakae-1 is running 0 daemon pod, expected 1 - Dec 14 17:05:19.824: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 - Dec 14 17:05:19.824: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set - STEP: Set a daemon pod's phase to 'Failed', check that the daemon pod is revived. 12/14/22 17:05:19.832 - Dec 14 17:05:19.880: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 - Dec 14 17:05:19.880: INFO: Node kiedai4iakae-3 is running 0 daemon pod, expected 1 - Dec 14 17:05:20.891: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 - Dec 14 17:05:20.891: INFO: Node kiedai4iakae-3 is running 0 daemon pod, expected 1 - Dec 14 17:05:21.896: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 2 - Dec 14 17:05:21.896: INFO: Node kiedai4iakae-3 is running 0 daemon pod, expected 1 - Dec 14 17:05:22.896: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 3 - Dec 14 17:05:22.896: INFO: Number of running nodes: 3, number of available pods: 3 in daemonset daemon-set - STEP: Wait for the failed daemon pod to be completely deleted. 12/14/22 17:05:22.896 - [AfterEach] [sig-apps] Daemon set [Serial] - test/e2e/apps/daemon_set.go:110 - STEP: Deleting DaemonSet "daemon-set" 12/14/22 17:05:22.906 - STEP: deleting DaemonSet.extensions daemon-set in namespace daemonsets-7350, will wait for the garbage collector to delete the pods 12/14/22 17:05:22.906 - Dec 14 17:05:22.970: INFO: Deleting DaemonSet.extensions daemon-set took: 8.007462ms - Dec 14 17:05:23.071: INFO: Terminating DaemonSet.extensions daemon-set pods took: 100.999927ms - Dec 14 17:05:25.782: INFO: Number of nodes with available pods controlled by daemonset daemon-set: 0 - Dec 14 17:05:25.782: INFO: Number of running nodes: 0, number of available pods: 0 in daemonset daemon-set - Dec 14 17:05:25.789: INFO: daemonset: {"kind":"DaemonSetList","apiVersion":"apps/v1","metadata":{"resourceVersion":"29734"},"items":null} - - Dec 14 17:05:25.800: INFO: pods: {"kind":"PodList","apiVersion":"v1","metadata":{"resourceVersion":"29734"},"items":null} - - [AfterEach] [sig-apps] Daemon set [Serial] + STEP: Creating a kubernetes client 02/25/23 11:31:26.056 + Feb 25 11:31:26.057: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-probe 02/25/23 11:31:26.059 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:31:26.082 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:31:26.086 + [BeforeEach] [sig-node] Probing container + test/e2e/common/node/container_probe.go:59 + [It] should *not* be restarted with a exec "cat /tmp/health" liveness probe [NodeConformance] [Conformance] + test/e2e/common/node/container_probe.go:148 + STEP: Creating pod busybox-de151162-692a-4290-832a-6281a820b040 in namespace container-probe-5837 02/25/23 11:31:26.09 + Feb 25 11:31:26.100: INFO: Waiting up to 5m0s for pod "busybox-de151162-692a-4290-832a-6281a820b040" in namespace "container-probe-5837" to be "not pending" + Feb 25 11:31:26.107: INFO: Pod "busybox-de151162-692a-4290-832a-6281a820b040": Phase="Pending", Reason="", readiness=false. Elapsed: 6.610336ms + Feb 25 11:31:28.112: INFO: Pod "busybox-de151162-692a-4290-832a-6281a820b040": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011684075s + Feb 25 11:31:30.113: INFO: Pod "busybox-de151162-692a-4290-832a-6281a820b040": Phase="Running", Reason="", readiness=true. Elapsed: 4.012813538s + Feb 25 11:31:30.113: INFO: Pod "busybox-de151162-692a-4290-832a-6281a820b040" satisfied condition "not pending" + Feb 25 11:31:30.113: INFO: Started pod busybox-de151162-692a-4290-832a-6281a820b040 in namespace container-probe-5837 + STEP: checking the pod's current state and verifying that restartCount is present 02/25/23 11:31:30.113 + Feb 25 11:31:30.118: INFO: Initial restart count of pod busybox-de151162-692a-4290-832a-6281a820b040 is 0 + STEP: deleting the pod 02/25/23 11:35:31.25 + [AfterEach] [sig-node] Probing container test/e2e/framework/framework.go:187 - Dec 14 17:05:25.819: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "daemonsets-7350" for this suite. 12/14/22 17:05:25.825 + Feb 25 11:35:31.285: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-probe-5837" for this suite. 02/25/23 11:35:31.296 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSS +SSSSSSSSSS ------------------------------ -[sig-storage] EmptyDir volumes - volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:156 -[BeforeEach] [sig-storage] EmptyDir volumes +[sig-node] Kubelet when scheduling a busybox command in a pod + should print the output to logs [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:52 +[BeforeEach] [sig-node] Kubelet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:05:25.84 -Dec 14 17:05:25.840: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir 12/14/22 17:05:25.841 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:25.893 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:25.896 -[It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:156 -STEP: Creating a pod to test emptydir volume type on node default medium 12/14/22 17:05:25.898 -Dec 14 17:05:25.912: INFO: Waiting up to 5m0s for pod "pod-a26db196-fdd3-481b-b7e3-c531ba262899" in namespace "emptydir-686" to be "Succeeded or Failed" -Dec 14 17:05:25.920: INFO: Pod "pod-a26db196-fdd3-481b-b7e3-c531ba262899": Phase="Pending", Reason="", readiness=false. Elapsed: 7.43934ms -Dec 14 17:05:27.925: INFO: Pod "pod-a26db196-fdd3-481b-b7e3-c531ba262899": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012792495s -Dec 14 17:05:29.926: INFO: Pod "pod-a26db196-fdd3-481b-b7e3-c531ba262899": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013220205s -STEP: Saw pod success 12/14/22 17:05:29.926 -Dec 14 17:05:29.926: INFO: Pod "pod-a26db196-fdd3-481b-b7e3-c531ba262899" satisfied condition "Succeeded or Failed" -Dec 14 17:05:29.929: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-a26db196-fdd3-481b-b7e3-c531ba262899 container test-container: -STEP: delete the pod 12/14/22 17:05:30.303 -Dec 14 17:05:30.328: INFO: Waiting for pod pod-a26db196-fdd3-481b-b7e3-c531ba262899 to disappear -Dec 14 17:05:30.340: INFO: Pod pod-a26db196-fdd3-481b-b7e3-c531ba262899 no longer exists -[AfterEach] [sig-storage] EmptyDir volumes +STEP: Creating a kubernetes client 02/25/23 11:35:31.321 +Feb 25 11:35:31.322: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubelet-test 02/25/23 11:35:31.325 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:35:31.37 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:35:31.374 +[BeforeEach] [sig-node] Kubelet + test/e2e/common/node/kubelet.go:41 +[It] should print the output to logs [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:52 +Feb 25 11:35:31.394: INFO: Waiting up to 5m0s for pod "busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39" in namespace "kubelet-test-6707" to be "running and ready" +Feb 25 11:35:31.406: INFO: Pod "busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39": Phase="Pending", Reason="", readiness=false. Elapsed: 11.945393ms +Feb 25 11:35:31.406: INFO: The phase of Pod busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:35:33.413: INFO: Pod "busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019306971s +Feb 25 11:35:33.414: INFO: The phase of Pod busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:35:35.415: INFO: Pod "busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39": Phase="Running", Reason="", readiness=true. Elapsed: 4.021060759s +Feb 25 11:35:35.415: INFO: The phase of Pod busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39 is Running (Ready = true) +Feb 25 11:35:35.415: INFO: Pod "busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39" satisfied condition "running and ready" +[AfterEach] [sig-node] Kubelet test/e2e/framework/framework.go:187 -Dec 14 17:05:30.340: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-686" for this suite. 12/14/22 17:05:30.354 -{"msg":"PASSED [sig-storage] EmptyDir volumes volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance]","completed":320,"skipped":6000,"failed":0} +Feb 25 11:35:35.452: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubelet-test-6707" for this suite. 02/25/23 11:35:35.461 +{"msg":"PASSED [sig-node] Kubelet when scheduling a busybox command in a pod should print the output to logs [NodeConformance] [Conformance]","completed":319,"skipped":6108,"failed":0} ------------------------------ -• [4.528 seconds] -[sig-storage] EmptyDir volumes -test/e2e/common/storage/framework.go:23 - volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:156 +• [4.170 seconds] +[sig-node] Kubelet +test/e2e/common/node/framework.go:23 + when scheduling a busybox command in a pod + test/e2e/common/node/kubelet.go:44 + should print the output to logs [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:52 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir volumes + [BeforeEach] [sig-node] Kubelet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:05:25.84 - Dec 14 17:05:25.840: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir 12/14/22 17:05:25.841 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:25.893 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:25.896 - [It] volume on default medium should have the correct mode [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/empty_dir.go:156 - STEP: Creating a pod to test emptydir volume type on node default medium 12/14/22 17:05:25.898 - Dec 14 17:05:25.912: INFO: Waiting up to 5m0s for pod "pod-a26db196-fdd3-481b-b7e3-c531ba262899" in namespace "emptydir-686" to be "Succeeded or Failed" - Dec 14 17:05:25.920: INFO: Pod "pod-a26db196-fdd3-481b-b7e3-c531ba262899": Phase="Pending", Reason="", readiness=false. Elapsed: 7.43934ms - Dec 14 17:05:27.925: INFO: Pod "pod-a26db196-fdd3-481b-b7e3-c531ba262899": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012792495s - Dec 14 17:05:29.926: INFO: Pod "pod-a26db196-fdd3-481b-b7e3-c531ba262899": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.013220205s - STEP: Saw pod success 12/14/22 17:05:29.926 - Dec 14 17:05:29.926: INFO: Pod "pod-a26db196-fdd3-481b-b7e3-c531ba262899" satisfied condition "Succeeded or Failed" - Dec 14 17:05:29.929: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-a26db196-fdd3-481b-b7e3-c531ba262899 container test-container: - STEP: delete the pod 12/14/22 17:05:30.303 - Dec 14 17:05:30.328: INFO: Waiting for pod pod-a26db196-fdd3-481b-b7e3-c531ba262899 to disappear - Dec 14 17:05:30.340: INFO: Pod pod-a26db196-fdd3-481b-b7e3-c531ba262899 no longer exists - [AfterEach] [sig-storage] EmptyDir volumes + STEP: Creating a kubernetes client 02/25/23 11:35:31.321 + Feb 25 11:35:31.322: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubelet-test 02/25/23 11:35:31.325 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:35:31.37 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:35:31.374 + [BeforeEach] [sig-node] Kubelet + test/e2e/common/node/kubelet.go:41 + [It] should print the output to logs [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:52 + Feb 25 11:35:31.394: INFO: Waiting up to 5m0s for pod "busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39" in namespace "kubelet-test-6707" to be "running and ready" + Feb 25 11:35:31.406: INFO: Pod "busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39": Phase="Pending", Reason="", readiness=false. Elapsed: 11.945393ms + Feb 25 11:35:31.406: INFO: The phase of Pod busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:35:33.413: INFO: Pod "busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39": Phase="Pending", Reason="", readiness=false. Elapsed: 2.019306971s + Feb 25 11:35:33.414: INFO: The phase of Pod busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:35:35.415: INFO: Pod "busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39": Phase="Running", Reason="", readiness=true. Elapsed: 4.021060759s + Feb 25 11:35:35.415: INFO: The phase of Pod busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39 is Running (Ready = true) + Feb 25 11:35:35.415: INFO: Pod "busybox-scheduling-2c95191b-e41f-462d-88af-2ea27570cc39" satisfied condition "running and ready" + [AfterEach] [sig-node] Kubelet test/e2e/framework/framework.go:187 - Dec 14 17:05:30.340: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-686" for this suite. 12/14/22 17:05:30.354 + Feb 25 11:35:35.452: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubelet-test-6707" for this suite. 02/25/23 11:35:35.461 << End Captured GinkgoWriter Output ------------------------------ -SSSSSS +SSSSSSSSSSS ------------------------------ -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - should be able to deny attaching pod [Conformance] - test/e2e/apimachinery/webhook.go:208 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +[sig-node] Containers + should be able to override the image's default arguments (container cmd) [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:58 +[BeforeEach] [sig-node] Containers test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:05:30.372 -Dec 14 17:05:30.373: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 17:05:30.383 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:30.451 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:30.456 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 17:05:30.481 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 17:05:31.855 -STEP: Deploying the webhook pod 12/14/22 17:05:31.881 -STEP: Wait for the deployment to be ready 12/14/22 17:05:31.904 -Dec 14 17:05:31.916: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set -STEP: Deploying the webhook service 12/14/22 17:05:33.93 -STEP: Verifying the service has paired with the endpoint 12/14/22 17:05:33.941 -Dec 14 17:05:34.942: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] should be able to deny attaching pod [Conformance] - test/e2e/apimachinery/webhook.go:208 -STEP: Registering the webhook via the AdmissionRegistration API 12/14/22 17:05:34.952 -STEP: create a pod 12/14/22 17:05:34.983 -Dec 14 17:05:34.995: INFO: Waiting up to 5m0s for pod "to-be-attached-pod" in namespace "webhook-1239" to be "running" -Dec 14 17:05:35.000: INFO: Pod "to-be-attached-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 5.403835ms -Dec 14 17:05:37.006: INFO: Pod "to-be-attached-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.011217391s -Dec 14 17:05:37.006: INFO: Pod "to-be-attached-pod" satisfied condition "running" -STEP: 'kubectl attach' the pod, should be denied by the webhook 12/14/22 17:05:37.006 -Dec 14 17:05:37.006: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=webhook-1239 attach --namespace=webhook-1239 to-be-attached-pod -i -c=container1' -Dec 14 17:05:37.207: INFO: rc: 1 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 11:35:35.496 +Feb 25 11:35:35.496: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename containers 02/25/23 11:35:35.498 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:35:35.525 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:35:35.53 +[It] should be able to override the image's default arguments (container cmd) [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:58 +STEP: Creating a pod to test override arguments 02/25/23 11:35:35.534 +Feb 25 11:35:35.551: INFO: Waiting up to 5m0s for pod "client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9" in namespace "containers-8024" to be "Succeeded or Failed" +Feb 25 11:35:35.556: INFO: Pod "client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9": Phase="Pending", Reason="", readiness=false. Elapsed: 5.312955ms +Feb 25 11:35:37.562: INFO: Pod "client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011526224s +Feb 25 11:35:39.564: INFO: Pod "client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012786594s +Feb 25 11:35:41.563: INFO: Pod "client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.012429891s +STEP: Saw pod success 02/25/23 11:35:41.564 +Feb 25 11:35:41.566: INFO: Pod "client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9" satisfied condition "Succeeded or Failed" +Feb 25 11:35:41.574: INFO: Trying to get logs from node diek3dua9gay-3 pod client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9 container agnhost-container: +STEP: delete the pod 02/25/23 11:35:41.584 +Feb 25 11:35:41.605: INFO: Waiting for pod client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9 to disappear +Feb 25 11:35:41.613: INFO: Pod client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9 no longer exists +[AfterEach] [sig-node] Containers test/e2e/framework/framework.go:187 -Dec 14 17:05:37.220: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-1239" for this suite. 12/14/22 17:05:37.226 -STEP: Destroying namespace "webhook-1239-markers" for this suite. 12/14/22 17:05:37.242 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny attaching pod [Conformance]","completed":321,"skipped":6006,"failed":0} +Feb 25 11:35:41.613: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "containers-8024" for this suite. 02/25/23 11:35:41.628 +{"msg":"PASSED [sig-node] Containers should be able to override the image's default arguments (container cmd) [NodeConformance] [Conformance]","completed":320,"skipped":6119,"failed":0} ------------------------------ -• [SLOW TEST] [6.943 seconds] -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - should be able to deny attaching pod [Conformance] - test/e2e/apimachinery/webhook.go:208 +• [SLOW TEST] [6.148 seconds] +[sig-node] Containers +test/e2e/common/node/framework.go:23 + should be able to override the image's default arguments (container cmd) [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:58 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + [BeforeEach] [sig-node] Containers test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:05:30.372 - Dec 14 17:05:30.373: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 17:05:30.383 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:30.451 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:30.456 - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 17:05:30.481 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 17:05:31.855 - STEP: Deploying the webhook pod 12/14/22 17:05:31.881 - STEP: Wait for the deployment to be ready 12/14/22 17:05:31.904 - Dec 14 17:05:31.916: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set - STEP: Deploying the webhook service 12/14/22 17:05:33.93 - STEP: Verifying the service has paired with the endpoint 12/14/22 17:05:33.941 - Dec 14 17:05:34.942: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] should be able to deny attaching pod [Conformance] - test/e2e/apimachinery/webhook.go:208 - STEP: Registering the webhook via the AdmissionRegistration API 12/14/22 17:05:34.952 - STEP: create a pod 12/14/22 17:05:34.983 - Dec 14 17:05:34.995: INFO: Waiting up to 5m0s for pod "to-be-attached-pod" in namespace "webhook-1239" to be "running" - Dec 14 17:05:35.000: INFO: Pod "to-be-attached-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 5.403835ms - Dec 14 17:05:37.006: INFO: Pod "to-be-attached-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.011217391s - Dec 14 17:05:37.006: INFO: Pod "to-be-attached-pod" satisfied condition "running" - STEP: 'kubectl attach' the pod, should be denied by the webhook 12/14/22 17:05:37.006 - Dec 14 17:05:37.006: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=webhook-1239 attach --namespace=webhook-1239 to-be-attached-pod -i -c=container1' - Dec 14 17:05:37.207: INFO: rc: 1 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 11:35:35.496 + Feb 25 11:35:35.496: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename containers 02/25/23 11:35:35.498 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:35:35.525 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:35:35.53 + [It] should be able to override the image's default arguments (container cmd) [NodeConformance] [Conformance] + test/e2e/common/node/containers.go:58 + STEP: Creating a pod to test override arguments 02/25/23 11:35:35.534 + Feb 25 11:35:35.551: INFO: Waiting up to 5m0s for pod "client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9" in namespace "containers-8024" to be "Succeeded or Failed" + Feb 25 11:35:35.556: INFO: Pod "client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9": Phase="Pending", Reason="", readiness=false. Elapsed: 5.312955ms + Feb 25 11:35:37.562: INFO: Pod "client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011526224s + Feb 25 11:35:39.564: INFO: Pod "client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012786594s + Feb 25 11:35:41.563: INFO: Pod "client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.012429891s + STEP: Saw pod success 02/25/23 11:35:41.564 + Feb 25 11:35:41.566: INFO: Pod "client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9" satisfied condition "Succeeded or Failed" + Feb 25 11:35:41.574: INFO: Trying to get logs from node diek3dua9gay-3 pod client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9 container agnhost-container: + STEP: delete the pod 02/25/23 11:35:41.584 + Feb 25 11:35:41.605: INFO: Waiting for pod client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9 to disappear + Feb 25 11:35:41.613: INFO: Pod client-containers-eb6b67cb-92b9-455e-909e-c0a833905bd9 no longer exists + [AfterEach] [sig-node] Containers test/e2e/framework/framework.go:187 - Dec 14 17:05:37.220: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-1239" for this suite. 12/14/22 17:05:37.226 - STEP: Destroying namespace "webhook-1239-markers" for this suite. 12/14/22 17:05:37.242 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 + Feb 25 11:35:41.613: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "containers-8024" for this suite. 02/25/23 11:35:41.628 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSS ------------------------------ -[sig-apps] ReplicaSet - Replace and Patch tests [Conformance] - test/e2e/apps/replica_set.go:154 -[BeforeEach] [sig-apps] ReplicaSet +[sig-node] RuntimeClass + should schedule a Pod requesting a RuntimeClass and initialize its Overhead [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:129 +[BeforeEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:05:37.318 -Dec 14 17:05:37.318: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename replicaset 12/14/22 17:05:37.321 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:37.38 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:37.383 -[It] Replace and Patch tests [Conformance] - test/e2e/apps/replica_set.go:154 -Dec 14 17:05:37.404: INFO: Pod name sample-pod: Found 0 pods out of 1 -Dec 14 17:05:42.410: INFO: Pod name sample-pod: Found 1 pods out of 1 -STEP: ensuring each pod is running 12/14/22 17:05:42.41 -STEP: Scaling up "test-rs" replicaset 12/14/22 17:05:42.41 -Dec 14 17:05:42.426: INFO: Updating replica set "test-rs" -STEP: patching the ReplicaSet 12/14/22 17:05:42.426 -W1214 17:05:42.439176 14 warnings.go:70] unknown field "spec.template.spec.TerminationGracePeriodSeconds" -Dec 14 17:05:42.440: INFO: observed ReplicaSet test-rs in namespace replicaset-5334 with ReadyReplicas 1, AvailableReplicas 1 -Dec 14 17:05:42.493: INFO: observed ReplicaSet test-rs in namespace replicaset-5334 with ReadyReplicas 1, AvailableReplicas 1 -Dec 14 17:05:42.566: INFO: observed ReplicaSet test-rs in namespace replicaset-5334 with ReadyReplicas 1, AvailableReplicas 1 -Dec 14 17:05:42.580: INFO: observed ReplicaSet test-rs in namespace replicaset-5334 with ReadyReplicas 1, AvailableReplicas 1 -Dec 14 17:05:43.671: INFO: observed ReplicaSet test-rs in namespace replicaset-5334 with ReadyReplicas 2, AvailableReplicas 2 -Dec 14 17:05:44.555: INFO: observed Replicaset test-rs in namespace replicaset-5334 with ReadyReplicas 3 found true -[AfterEach] [sig-apps] ReplicaSet +STEP: Creating a kubernetes client 02/25/23 11:35:41.653 +Feb 25 11:35:41.653: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename runtimeclass 02/25/23 11:35:41.666 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:35:41.706 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:35:41.711 +[It] should schedule a Pod requesting a RuntimeClass and initialize its Overhead [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:129 +Feb 25 11:35:41.743: INFO: Waiting up to 1m20s for at least 1 pods in namespace runtimeclass-9669 to be scheduled +Feb 25 11:35:41.748: INFO: 1 pods are not scheduled: [runtimeclass-9669/test-runtimeclass-runtimeclass-9669-preconfigured-handler-j486r(d7d48f6c-0ed5-48f3-abed-5e1b1106a25a)] +[AfterEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:187 -Dec 14 17:05:44.555: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "replicaset-5334" for this suite. 12/14/22 17:05:44.562 -{"msg":"PASSED [sig-apps] ReplicaSet Replace and Patch tests [Conformance]","completed":322,"skipped":6037,"failed":0} +Feb 25 11:35:43.764: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "runtimeclass-9669" for this suite. 02/25/23 11:35:43.771 +{"msg":"PASSED [sig-node] RuntimeClass should schedule a Pod requesting a RuntimeClass and initialize its Overhead [NodeConformance] [Conformance]","completed":321,"skipped":6123,"failed":0} ------------------------------ -• [SLOW TEST] [7.252 seconds] -[sig-apps] ReplicaSet -test/e2e/apps/framework.go:23 - Replace and Patch tests [Conformance] - test/e2e/apps/replica_set.go:154 +• [2.128 seconds] +[sig-node] RuntimeClass +test/e2e/common/node/framework.go:23 + should schedule a Pod requesting a RuntimeClass and initialize its Overhead [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:129 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] ReplicaSet + [BeforeEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:05:37.318 - Dec 14 17:05:37.318: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename replicaset 12/14/22 17:05:37.321 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:37.38 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:37.383 - [It] Replace and Patch tests [Conformance] - test/e2e/apps/replica_set.go:154 - Dec 14 17:05:37.404: INFO: Pod name sample-pod: Found 0 pods out of 1 - Dec 14 17:05:42.410: INFO: Pod name sample-pod: Found 1 pods out of 1 - STEP: ensuring each pod is running 12/14/22 17:05:42.41 - STEP: Scaling up "test-rs" replicaset 12/14/22 17:05:42.41 - Dec 14 17:05:42.426: INFO: Updating replica set "test-rs" - STEP: patching the ReplicaSet 12/14/22 17:05:42.426 - W1214 17:05:42.439176 14 warnings.go:70] unknown field "spec.template.spec.TerminationGracePeriodSeconds" - Dec 14 17:05:42.440: INFO: observed ReplicaSet test-rs in namespace replicaset-5334 with ReadyReplicas 1, AvailableReplicas 1 - Dec 14 17:05:42.493: INFO: observed ReplicaSet test-rs in namespace replicaset-5334 with ReadyReplicas 1, AvailableReplicas 1 - Dec 14 17:05:42.566: INFO: observed ReplicaSet test-rs in namespace replicaset-5334 with ReadyReplicas 1, AvailableReplicas 1 - Dec 14 17:05:42.580: INFO: observed ReplicaSet test-rs in namespace replicaset-5334 with ReadyReplicas 1, AvailableReplicas 1 - Dec 14 17:05:43.671: INFO: observed ReplicaSet test-rs in namespace replicaset-5334 with ReadyReplicas 2, AvailableReplicas 2 - Dec 14 17:05:44.555: INFO: observed Replicaset test-rs in namespace replicaset-5334 with ReadyReplicas 3 found true - [AfterEach] [sig-apps] ReplicaSet + STEP: Creating a kubernetes client 02/25/23 11:35:41.653 + Feb 25 11:35:41.653: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename runtimeclass 02/25/23 11:35:41.666 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:35:41.706 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:35:41.711 + [It] should schedule a Pod requesting a RuntimeClass and initialize its Overhead [NodeConformance] [Conformance] + test/e2e/common/node/runtimeclass.go:129 + Feb 25 11:35:41.743: INFO: Waiting up to 1m20s for at least 1 pods in namespace runtimeclass-9669 to be scheduled + Feb 25 11:35:41.748: INFO: 1 pods are not scheduled: [runtimeclass-9669/test-runtimeclass-runtimeclass-9669-preconfigured-handler-j486r(d7d48f6c-0ed5-48f3-abed-5e1b1106a25a)] + [AfterEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:187 - Dec 14 17:05:44.555: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "replicaset-5334" for this suite. 12/14/22 17:05:44.562 + Feb 25 11:35:43.764: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "runtimeclass-9669" for this suite. 02/25/23 11:35:43.771 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSS +S ------------------------------ -[sig-storage] Projected downwardAPI - should provide podname only [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:52 -[BeforeEach] [sig-storage] Projected downwardAPI +[sig-node] Pods + should be updated [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:343 +[BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:05:44.572 -Dec 14 17:05:44.572: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 17:05:44.575 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:44.597 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:44.602 -[BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 -[It] should provide podname only [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:52 -STEP: Creating a pod to test downward API volume plugin 12/14/22 17:05:44.605 -Dec 14 17:05:44.619: INFO: Waiting up to 5m0s for pod "downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf" in namespace "projected-5451" to be "Succeeded or Failed" -Dec 14 17:05:44.623: INFO: Pod "downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf": Phase="Pending", Reason="", readiness=false. Elapsed: 4.199629ms -Dec 14 17:05:46.629: INFO: Pod "downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf": Phase="Running", Reason="", readiness=true. Elapsed: 2.009656766s -Dec 14 17:05:48.631: INFO: Pod "downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf": Phase="Running", Reason="", readiness=false. Elapsed: 4.012417654s -Dec 14 17:05:50.637: INFO: Pod "downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.017577146s -STEP: Saw pod success 12/14/22 17:05:50.637 -Dec 14 17:05:50.637: INFO: Pod "downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf" satisfied condition "Succeeded or Failed" -Dec 14 17:05:50.642: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf container client-container: -STEP: delete the pod 12/14/22 17:05:50.654 -Dec 14 17:05:50.670: INFO: Waiting for pod downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf to disappear -Dec 14 17:05:50.680: INFO: Pod downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf no longer exists -[AfterEach] [sig-storage] Projected downwardAPI +STEP: Creating a kubernetes client 02/25/23 11:35:43.781 +Feb 25 11:35:43.781: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename pods 02/25/23 11:35:43.783 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:35:43.806 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:35:43.81 +[BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 +[It] should be updated [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:343 +STEP: creating the pod 02/25/23 11:35:43.815 +STEP: submitting the pod to kubernetes 02/25/23 11:35:43.815 +Feb 25 11:35:43.829: INFO: Waiting up to 5m0s for pod "pod-update-7e358426-a428-4e77-be82-2037567fc888" in namespace "pods-2500" to be "running and ready" +Feb 25 11:35:43.835: INFO: Pod "pod-update-7e358426-a428-4e77-be82-2037567fc888": Phase="Pending", Reason="", readiness=false. Elapsed: 6.186581ms +Feb 25 11:35:43.835: INFO: The phase of Pod pod-update-7e358426-a428-4e77-be82-2037567fc888 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:35:45.843: INFO: Pod "pod-update-7e358426-a428-4e77-be82-2037567fc888": Phase="Running", Reason="", readiness=true. Elapsed: 2.013610982s +Feb 25 11:35:45.843: INFO: The phase of Pod pod-update-7e358426-a428-4e77-be82-2037567fc888 is Running (Ready = true) +Feb 25 11:35:45.843: INFO: Pod "pod-update-7e358426-a428-4e77-be82-2037567fc888" satisfied condition "running and ready" +STEP: verifying the pod is in kubernetes 02/25/23 11:35:45.847 +STEP: updating the pod 02/25/23 11:35:45.852 +Feb 25 11:35:46.371: INFO: Successfully updated pod "pod-update-7e358426-a428-4e77-be82-2037567fc888" +Feb 25 11:35:46.371: INFO: Waiting up to 5m0s for pod "pod-update-7e358426-a428-4e77-be82-2037567fc888" in namespace "pods-2500" to be "running" +Feb 25 11:35:46.383: INFO: Pod "pod-update-7e358426-a428-4e77-be82-2037567fc888": Phase="Running", Reason="", readiness=true. Elapsed: 11.271391ms +Feb 25 11:35:46.383: INFO: Pod "pod-update-7e358426-a428-4e77-be82-2037567fc888" satisfied condition "running" +STEP: verifying the updated pod is in kubernetes 02/25/23 11:35:46.383 +Feb 25 11:35:46.400: INFO: Pod update OK +[AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 -Dec 14 17:05:50.680: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-5451" for this suite. 12/14/22 17:05:50.687 -{"msg":"PASSED [sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance]","completed":323,"skipped":6049,"failed":0} +Feb 25 11:35:46.400: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pods-2500" for this suite. 02/25/23 11:35:46.411 +{"msg":"PASSED [sig-node] Pods should be updated [NodeConformance] [Conformance]","completed":322,"skipped":6124,"failed":0} ------------------------------ -• [SLOW TEST] [6.127 seconds] -[sig-storage] Projected downwardAPI -test/e2e/common/storage/framework.go:23 - should provide podname only [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:52 +• [2.641 seconds] +[sig-node] Pods +test/e2e/common/node/framework.go:23 + should be updated [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:343 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected downwardAPI + [BeforeEach] [sig-node] Pods test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:05:44.572 - Dec 14 17:05:44.572: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 17:05:44.575 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:44.597 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:44.602 - [BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 - [It] should provide podname only [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:52 - STEP: Creating a pod to test downward API volume plugin 12/14/22 17:05:44.605 - Dec 14 17:05:44.619: INFO: Waiting up to 5m0s for pod "downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf" in namespace "projected-5451" to be "Succeeded or Failed" - Dec 14 17:05:44.623: INFO: Pod "downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf": Phase="Pending", Reason="", readiness=false. Elapsed: 4.199629ms - Dec 14 17:05:46.629: INFO: Pod "downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf": Phase="Running", Reason="", readiness=true. Elapsed: 2.009656766s - Dec 14 17:05:48.631: INFO: Pod "downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf": Phase="Running", Reason="", readiness=false. Elapsed: 4.012417654s - Dec 14 17:05:50.637: INFO: Pod "downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.017577146s - STEP: Saw pod success 12/14/22 17:05:50.637 - Dec 14 17:05:50.637: INFO: Pod "downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf" satisfied condition "Succeeded or Failed" - Dec 14 17:05:50.642: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf container client-container: - STEP: delete the pod 12/14/22 17:05:50.654 - Dec 14 17:05:50.670: INFO: Waiting for pod downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf to disappear - Dec 14 17:05:50.680: INFO: Pod downwardapi-volume-fa477852-7c24-4494-a981-b730705626cf no longer exists - [AfterEach] [sig-storage] Projected downwardAPI + STEP: Creating a kubernetes client 02/25/23 11:35:43.781 + Feb 25 11:35:43.781: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename pods 02/25/23 11:35:43.783 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:35:43.806 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:35:43.81 + [BeforeEach] [sig-node] Pods + test/e2e/common/node/pods.go:193 + [It] should be updated [NodeConformance] [Conformance] + test/e2e/common/node/pods.go:343 + STEP: creating the pod 02/25/23 11:35:43.815 + STEP: submitting the pod to kubernetes 02/25/23 11:35:43.815 + Feb 25 11:35:43.829: INFO: Waiting up to 5m0s for pod "pod-update-7e358426-a428-4e77-be82-2037567fc888" in namespace "pods-2500" to be "running and ready" + Feb 25 11:35:43.835: INFO: Pod "pod-update-7e358426-a428-4e77-be82-2037567fc888": Phase="Pending", Reason="", readiness=false. Elapsed: 6.186581ms + Feb 25 11:35:43.835: INFO: The phase of Pod pod-update-7e358426-a428-4e77-be82-2037567fc888 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:35:45.843: INFO: Pod "pod-update-7e358426-a428-4e77-be82-2037567fc888": Phase="Running", Reason="", readiness=true. Elapsed: 2.013610982s + Feb 25 11:35:45.843: INFO: The phase of Pod pod-update-7e358426-a428-4e77-be82-2037567fc888 is Running (Ready = true) + Feb 25 11:35:45.843: INFO: Pod "pod-update-7e358426-a428-4e77-be82-2037567fc888" satisfied condition "running and ready" + STEP: verifying the pod is in kubernetes 02/25/23 11:35:45.847 + STEP: updating the pod 02/25/23 11:35:45.852 + Feb 25 11:35:46.371: INFO: Successfully updated pod "pod-update-7e358426-a428-4e77-be82-2037567fc888" + Feb 25 11:35:46.371: INFO: Waiting up to 5m0s for pod "pod-update-7e358426-a428-4e77-be82-2037567fc888" in namespace "pods-2500" to be "running" + Feb 25 11:35:46.383: INFO: Pod "pod-update-7e358426-a428-4e77-be82-2037567fc888": Phase="Running", Reason="", readiness=true. Elapsed: 11.271391ms + Feb 25 11:35:46.383: INFO: Pod "pod-update-7e358426-a428-4e77-be82-2037567fc888" satisfied condition "running" + STEP: verifying the updated pod is in kubernetes 02/25/23 11:35:46.383 + Feb 25 11:35:46.400: INFO: Pod update OK + [AfterEach] [sig-node] Pods test/e2e/framework/framework.go:187 - Dec 14 17:05:50.680: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-5451" for this suite. 12/14/22 17:05:50.687 + Feb 25 11:35:46.400: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "pods-2500" for this suite. 02/25/23 11:35:46.411 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSS +SSSSSSSSSSS ------------------------------ -[sig-api-machinery] Namespaces [Serial] - should patch a Namespace [Conformance] - test/e2e/apimachinery/namespace.go:267 -[BeforeEach] [sig-api-machinery] Namespaces [Serial] +[sig-network] Ingress API + should support creating Ingress API operations [Conformance] + test/e2e/network/ingress.go:552 +[BeforeEach] [sig-network] Ingress API test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:05:50.703 -Dec 14 17:05:50.703: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename namespaces 12/14/22 17:05:50.706 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:50.731 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:50.734 -[It] should patch a Namespace [Conformance] - test/e2e/apimachinery/namespace.go:267 -STEP: creating a Namespace 12/14/22 17:05:50.738 -STEP: patching the Namespace 12/14/22 17:05:50.77 -STEP: get the Namespace and ensuring it has the label 12/14/22 17:05:50.777 -[AfterEach] [sig-api-machinery] Namespaces [Serial] +STEP: Creating a kubernetes client 02/25/23 11:35:46.425 +Feb 25 11:35:46.425: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename ingress 02/25/23 11:35:46.431 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:35:46.461 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:35:46.466 +[It] should support creating Ingress API operations [Conformance] + test/e2e/network/ingress.go:552 +STEP: getting /apis 02/25/23 11:35:46.47 +STEP: getting /apis/networking.k8s.io 02/25/23 11:35:46.472 +STEP: getting /apis/networking.k8s.iov1 02/25/23 11:35:46.474 +STEP: creating 02/25/23 11:35:46.475 +STEP: getting 02/25/23 11:35:46.499 +STEP: listing 02/25/23 11:35:46.503 +STEP: watching 02/25/23 11:35:46.516 +Feb 25 11:35:46.516: INFO: starting watch +STEP: cluster-wide listing 02/25/23 11:35:46.518 +STEP: cluster-wide watching 02/25/23 11:35:46.524 +Feb 25 11:35:46.524: INFO: starting watch +STEP: patching 02/25/23 11:35:46.525 +STEP: updating 02/25/23 11:35:46.532 +Feb 25 11:35:46.545: INFO: waiting for watch events with expected annotations +Feb 25 11:35:46.545: INFO: saw patched and updated annotations +STEP: patching /status 02/25/23 11:35:46.545 +STEP: updating /status 02/25/23 11:35:46.553 +STEP: get /status 02/25/23 11:35:46.567 +STEP: deleting 02/25/23 11:35:46.572 +STEP: deleting a collection 02/25/23 11:35:46.593 +[AfterEach] [sig-network] Ingress API test/e2e/framework/framework.go:187 -Dec 14 17:05:50.784: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "namespaces-6284" for this suite. 12/14/22 17:05:50.793 -STEP: Destroying namespace "nspatchtest-9821b3d9-c0f0-4477-8529-bf15c62a2606-9580" for this suite. 12/14/22 17:05:50.807 -{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should patch a Namespace [Conformance]","completed":324,"skipped":6063,"failed":0} +Feb 25 11:35:46.621: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "ingress-3038" for this suite. 02/25/23 11:35:46.633 +{"msg":"PASSED [sig-network] Ingress API should support creating Ingress API operations [Conformance]","completed":323,"skipped":6135,"failed":0} ------------------------------ -• [0.121 seconds] -[sig-api-machinery] Namespaces [Serial] -test/e2e/apimachinery/framework.go:23 - should patch a Namespace [Conformance] - test/e2e/apimachinery/namespace.go:267 +• [0.222 seconds] +[sig-network] Ingress API +test/e2e/network/common/framework.go:23 + should support creating Ingress API operations [Conformance] + test/e2e/network/ingress.go:552 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] Namespaces [Serial] + [BeforeEach] [sig-network] Ingress API test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:05:50.703 - Dec 14 17:05:50.703: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename namespaces 12/14/22 17:05:50.706 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:50.731 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:50.734 - [It] should patch a Namespace [Conformance] - test/e2e/apimachinery/namespace.go:267 - STEP: creating a Namespace 12/14/22 17:05:50.738 - STEP: patching the Namespace 12/14/22 17:05:50.77 - STEP: get the Namespace and ensuring it has the label 12/14/22 17:05:50.777 - [AfterEach] [sig-api-machinery] Namespaces [Serial] + STEP: Creating a kubernetes client 02/25/23 11:35:46.425 + Feb 25 11:35:46.425: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename ingress 02/25/23 11:35:46.431 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:35:46.461 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:35:46.466 + [It] should support creating Ingress API operations [Conformance] + test/e2e/network/ingress.go:552 + STEP: getting /apis 02/25/23 11:35:46.47 + STEP: getting /apis/networking.k8s.io 02/25/23 11:35:46.472 + STEP: getting /apis/networking.k8s.iov1 02/25/23 11:35:46.474 + STEP: creating 02/25/23 11:35:46.475 + STEP: getting 02/25/23 11:35:46.499 + STEP: listing 02/25/23 11:35:46.503 + STEP: watching 02/25/23 11:35:46.516 + Feb 25 11:35:46.516: INFO: starting watch + STEP: cluster-wide listing 02/25/23 11:35:46.518 + STEP: cluster-wide watching 02/25/23 11:35:46.524 + Feb 25 11:35:46.524: INFO: starting watch + STEP: patching 02/25/23 11:35:46.525 + STEP: updating 02/25/23 11:35:46.532 + Feb 25 11:35:46.545: INFO: waiting for watch events with expected annotations + Feb 25 11:35:46.545: INFO: saw patched and updated annotations + STEP: patching /status 02/25/23 11:35:46.545 + STEP: updating /status 02/25/23 11:35:46.553 + STEP: get /status 02/25/23 11:35:46.567 + STEP: deleting 02/25/23 11:35:46.572 + STEP: deleting a collection 02/25/23 11:35:46.593 + [AfterEach] [sig-network] Ingress API test/e2e/framework/framework.go:187 - Dec 14 17:05:50.784: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "namespaces-6284" for this suite. 12/14/22 17:05:50.793 - STEP: Destroying namespace "nspatchtest-9821b3d9-c0f0-4477-8529-bf15c62a2606-9580" for this suite. 12/14/22 17:05:50.807 + Feb 25 11:35:46.621: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "ingress-3038" for this suite. 02/25/23 11:35:46.633 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSS +SS ------------------------------ -[sig-node] Probing container - should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:165 -[BeforeEach] [sig-node] Probing container +[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] + should be able to convert a non homogeneous list of CRs [Conformance] + test/e2e/apimachinery/crd_conversion_webhook.go:184 +[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:05:50.832 -Dec 14 17:05:50.832: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-probe 12/14/22 17:05:50.836 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:50.885 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:50.889 -[BeforeEach] [sig-node] Probing container - test/e2e/common/node/container_probe.go:59 -[It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:165 -STEP: Creating pod liveness-671b9f89-4f42-424d-8d77-d5943c76d50d in namespace container-probe-5767 12/14/22 17:05:50.892 -Dec 14 17:05:50.903: INFO: Waiting up to 5m0s for pod "liveness-671b9f89-4f42-424d-8d77-d5943c76d50d" in namespace "container-probe-5767" to be "not pending" -Dec 14 17:05:50.906: INFO: Pod "liveness-671b9f89-4f42-424d-8d77-d5943c76d50d": Phase="Pending", Reason="", readiness=false. Elapsed: 3.464053ms -Dec 14 17:05:52.915: INFO: Pod "liveness-671b9f89-4f42-424d-8d77-d5943c76d50d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011953576s -Dec 14 17:05:54.912: INFO: Pod "liveness-671b9f89-4f42-424d-8d77-d5943c76d50d": Phase="Running", Reason="", readiness=true. Elapsed: 4.009037746s -Dec 14 17:05:54.912: INFO: Pod "liveness-671b9f89-4f42-424d-8d77-d5943c76d50d" satisfied condition "not pending" -Dec 14 17:05:54.912: INFO: Started pod liveness-671b9f89-4f42-424d-8d77-d5943c76d50d in namespace container-probe-5767 -STEP: checking the pod's current state and verifying that restartCount is present 12/14/22 17:05:54.912 -Dec 14 17:05:54.916: INFO: Initial restart count of pod liveness-671b9f89-4f42-424d-8d77-d5943c76d50d is 0 -Dec 14 17:06:14.998: INFO: Restart count of pod container-probe-5767/liveness-671b9f89-4f42-424d-8d77-d5943c76d50d is now 1 (20.082098741s elapsed) -STEP: deleting the pod 12/14/22 17:06:14.999 -[AfterEach] [sig-node] Probing container +STEP: Creating a kubernetes client 02/25/23 11:35:46.647 +Feb 25 11:35:46.648: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename crd-webhook 02/25/23 11:35:46.65 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:35:46.682 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:35:46.687 +[BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/crd_conversion_webhook.go:128 +STEP: Setting up server cert 02/25/23 11:35:46.691 +STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication 02/25/23 11:35:47.755 +STEP: Deploying the custom resource conversion webhook pod 02/25/23 11:35:47.766 +STEP: Wait for the deployment to be ready 02/25/23 11:35:47.794 +Feb 25 11:35:47.805: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set +Feb 25 11:35:49.825: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 11, 35, 47, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 35, 47, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 11, 35, 47, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 35, 47, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-59dfc5db8d\" is progressing."}}, CollisionCount:(*int32)(nil)} +STEP: Deploying the webhook service 02/25/23 11:35:51.833 +STEP: Verifying the service has paired with the endpoint 02/25/23 11:35:51.858 +Feb 25 11:35:52.858: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1 +[It] should be able to convert a non homogeneous list of CRs [Conformance] + test/e2e/apimachinery/crd_conversion_webhook.go:184 +Feb 25 11:35:52.867: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Creating a v1 custom resource 02/25/23 11:35:55.689 +STEP: Create a v2 custom resource 02/25/23 11:35:55.739 +STEP: List CRs in v1 02/25/23 11:35:55.765 +STEP: List CRs in v2 02/25/23 11:35:55.957 +[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 17:06:15.037: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-probe-5767" for this suite. 12/14/22 17:06:15.047 -{"msg":"PASSED [sig-node] Probing container should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance]","completed":325,"skipped":6074,"failed":0} +Feb 25 11:35:56.487: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "crd-webhook-7498" for this suite. 02/25/23 11:35:56.494 +[AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/crd_conversion_webhook.go:139 +{"msg":"PASSED [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] should be able to convert a non homogeneous list of CRs [Conformance]","completed":324,"skipped":6137,"failed":0} ------------------------------ -• [SLOW TEST] [24.225 seconds] -[sig-node] Probing container -test/e2e/common/node/framework.go:23 - should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:165 +• [SLOW TEST] [9.920 seconds] +[sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + should be able to convert a non homogeneous list of CRs [Conformance] + test/e2e/apimachinery/crd_conversion_webhook.go:184 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Probing container + [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:05:50.832 - Dec 14 17:05:50.832: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-probe 12/14/22 17:05:50.836 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:05:50.885 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:05:50.889 - [BeforeEach] [sig-node] Probing container - test/e2e/common/node/container_probe.go:59 - [It] should be restarted with a /healthz http liveness probe [NodeConformance] [Conformance] - test/e2e/common/node/container_probe.go:165 - STEP: Creating pod liveness-671b9f89-4f42-424d-8d77-d5943c76d50d in namespace container-probe-5767 12/14/22 17:05:50.892 - Dec 14 17:05:50.903: INFO: Waiting up to 5m0s for pod "liveness-671b9f89-4f42-424d-8d77-d5943c76d50d" in namespace "container-probe-5767" to be "not pending" - Dec 14 17:05:50.906: INFO: Pod "liveness-671b9f89-4f42-424d-8d77-d5943c76d50d": Phase="Pending", Reason="", readiness=false. Elapsed: 3.464053ms - Dec 14 17:05:52.915: INFO: Pod "liveness-671b9f89-4f42-424d-8d77-d5943c76d50d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011953576s - Dec 14 17:05:54.912: INFO: Pod "liveness-671b9f89-4f42-424d-8d77-d5943c76d50d": Phase="Running", Reason="", readiness=true. Elapsed: 4.009037746s - Dec 14 17:05:54.912: INFO: Pod "liveness-671b9f89-4f42-424d-8d77-d5943c76d50d" satisfied condition "not pending" - Dec 14 17:05:54.912: INFO: Started pod liveness-671b9f89-4f42-424d-8d77-d5943c76d50d in namespace container-probe-5767 - STEP: checking the pod's current state and verifying that restartCount is present 12/14/22 17:05:54.912 - Dec 14 17:05:54.916: INFO: Initial restart count of pod liveness-671b9f89-4f42-424d-8d77-d5943c76d50d is 0 - Dec 14 17:06:14.998: INFO: Restart count of pod container-probe-5767/liveness-671b9f89-4f42-424d-8d77-d5943c76d50d is now 1 (20.082098741s elapsed) - STEP: deleting the pod 12/14/22 17:06:14.999 - [AfterEach] [sig-node] Probing container + STEP: Creating a kubernetes client 02/25/23 11:35:46.647 + Feb 25 11:35:46.648: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename crd-webhook 02/25/23 11:35:46.65 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:35:46.682 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:35:46.687 + [BeforeEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/crd_conversion_webhook.go:128 + STEP: Setting up server cert 02/25/23 11:35:46.691 + STEP: Create role binding to let cr conversion webhook read extension-apiserver-authentication 02/25/23 11:35:47.755 + STEP: Deploying the custom resource conversion webhook pod 02/25/23 11:35:47.766 + STEP: Wait for the deployment to be ready 02/25/23 11:35:47.794 + Feb 25 11:35:47.805: INFO: deployment "sample-crd-conversion-webhook-deployment" doesn't have the required revision set + Feb 25 11:35:49.825: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 11, 35, 47, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 35, 47, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 11, 35, 47, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 35, 47, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-crd-conversion-webhook-deployment-59dfc5db8d\" is progressing."}}, CollisionCount:(*int32)(nil)} + STEP: Deploying the webhook service 02/25/23 11:35:51.833 + STEP: Verifying the service has paired with the endpoint 02/25/23 11:35:51.858 + Feb 25 11:35:52.858: INFO: Waiting for amount of service:e2e-test-crd-conversion-webhook endpoints to be 1 + [It] should be able to convert a non homogeneous list of CRs [Conformance] + test/e2e/apimachinery/crd_conversion_webhook.go:184 + Feb 25 11:35:52.867: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Creating a v1 custom resource 02/25/23 11:35:55.689 + STEP: Create a v2 custom resource 02/25/23 11:35:55.739 + STEP: List CRs in v1 02/25/23 11:35:55.765 + STEP: List CRs in v2 02/25/23 11:35:55.957 + [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 17:06:15.037: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-probe-5767" for this suite. 12/14/22 17:06:15.047 + Feb 25 11:35:56.487: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "crd-webhook-7498" for this suite. 02/25/23 11:35:56.494 + [AfterEach] [sig-api-machinery] CustomResourceConversionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/crd_conversion_webhook.go:139 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSS +SSS ------------------------------ -[sig-apps] CronJob - should support CronJob API operations [Conformance] - test/e2e/apps/cronjob.go:319 -[BeforeEach] [sig-apps] CronJob +[sig-network] Services + should serve multiport endpoints from pods [Conformance] + test/e2e/network/service.go:852 +[BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:06:15.065 -Dec 14 17:06:15.065: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename cronjob 12/14/22 17:06:15.066 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:15.092 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:15.095 -[It] should support CronJob API operations [Conformance] - test/e2e/apps/cronjob.go:319 -STEP: Creating a cronjob 12/14/22 17:06:15.099 -STEP: creating 12/14/22 17:06:15.099 -STEP: getting 12/14/22 17:06:15.111 -STEP: listing 12/14/22 17:06:15.114 -STEP: watching 12/14/22 17:06:15.121 -Dec 14 17:06:15.121: INFO: starting watch -STEP: cluster-wide listing 12/14/22 17:06:15.131 -STEP: cluster-wide watching 12/14/22 17:06:15.137 -Dec 14 17:06:15.137: INFO: starting watch -STEP: patching 12/14/22 17:06:15.138 -STEP: updating 12/14/22 17:06:15.161 -Dec 14 17:06:15.188: INFO: waiting for watch events with expected annotations -Dec 14 17:06:15.189: INFO: saw patched and updated annotations -STEP: patching /status 12/14/22 17:06:15.189 -STEP: updating /status 12/14/22 17:06:15.209 -STEP: get /status 12/14/22 17:06:15.225 -STEP: deleting 12/14/22 17:06:15.228 -STEP: deleting a collection 12/14/22 17:06:15.252 -[AfterEach] [sig-apps] CronJob +STEP: Creating a kubernetes client 02/25/23 11:35:56.568 +Feb 25 11:35:56.568: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 11:35:56.57 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:35:56.677 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:35:56.681 +[BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 +[It] should serve multiport endpoints from pods [Conformance] + test/e2e/network/service.go:852 +STEP: creating service multi-endpoint-test in namespace services-9861 02/25/23 11:35:56.689 +STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-9861 to expose endpoints map[] 02/25/23 11:35:56.749 +Feb 25 11:35:56.773: INFO: successfully validated that service multi-endpoint-test in namespace services-9861 exposes endpoints map[] +STEP: Creating pod pod1 in namespace services-9861 02/25/23 11:35:56.773 +Feb 25 11:35:56.792: INFO: Waiting up to 5m0s for pod "pod1" in namespace "services-9861" to be "running and ready" +Feb 25 11:35:56.822: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 29.691884ms +Feb 25 11:35:56.822: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:35:58.830: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.037419176s +Feb 25 11:35:58.830: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:36:00.831: INFO: Pod "pod1": Phase="Running", Reason="", readiness=true. Elapsed: 4.038115024s +Feb 25 11:36:00.831: INFO: The phase of Pod pod1 is Running (Ready = true) +Feb 25 11:36:00.831: INFO: Pod "pod1" satisfied condition "running and ready" +STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-9861 to expose endpoints map[pod1:[100]] 02/25/23 11:36:00.836 +Feb 25 11:36:00.851: INFO: successfully validated that service multi-endpoint-test in namespace services-9861 exposes endpoints map[pod1:[100]] +STEP: Creating pod pod2 in namespace services-9861 02/25/23 11:36:00.852 +Feb 25 11:36:00.921: INFO: Waiting up to 5m0s for pod "pod2" in namespace "services-9861" to be "running and ready" +Feb 25 11:36:00.955: INFO: Pod "pod2": Phase="Pending", Reason="", readiness=false. Elapsed: 34.080947ms +Feb 25 11:36:00.955: INFO: The phase of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:36:02.963: INFO: Pod "pod2": Phase="Running", Reason="", readiness=true. Elapsed: 2.042312421s +Feb 25 11:36:02.964: INFO: The phase of Pod pod2 is Running (Ready = true) +Feb 25 11:36:02.964: INFO: Pod "pod2" satisfied condition "running and ready" +STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-9861 to expose endpoints map[pod1:[100] pod2:[101]] 02/25/23 11:36:02.977 +Feb 25 11:36:03.010: INFO: successfully validated that service multi-endpoint-test in namespace services-9861 exposes endpoints map[pod1:[100] pod2:[101]] +STEP: Checking if the Service forwards traffic to pods 02/25/23 11:36:03.01 +Feb 25 11:36:03.011: INFO: Creating new exec pod +Feb 25 11:36:03.027: INFO: Waiting up to 5m0s for pod "execpod2k2h4" in namespace "services-9861" to be "running" +Feb 25 11:36:03.038: INFO: Pod "execpod2k2h4": Phase="Pending", Reason="", readiness=false. Elapsed: 11.313781ms +Feb 25 11:36:05.049: INFO: Pod "execpod2k2h4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02216235s +Feb 25 11:36:07.046: INFO: Pod "execpod2k2h4": Phase="Running", Reason="", readiness=true. Elapsed: 4.01866938s +Feb 25 11:36:07.046: INFO: Pod "execpod2k2h4" satisfied condition "running" +Feb 25 11:36:08.047: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9861 exec execpod2k2h4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 multi-endpoint-test 80' +Feb 25 11:36:08.411: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 multi-endpoint-test 80\nConnection to multi-endpoint-test 80 port [tcp/http] succeeded!\n" +Feb 25 11:36:08.411: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:36:08.411: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9861 exec execpod2k2h4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.38.253 80' +Feb 25 11:36:08.733: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.38.253 80\nConnection to 10.233.38.253 80 port [tcp/http] succeeded!\n" +Feb 25 11:36:08.733: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:36:08.733: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9861 exec execpod2k2h4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 multi-endpoint-test 81' +Feb 25 11:36:09.165: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 multi-endpoint-test 81\nConnection to multi-endpoint-test 81 port [tcp/*] succeeded!\n" +Feb 25 11:36:09.165: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +Feb 25 11:36:09.165: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9861 exec execpod2k2h4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.38.253 81' +Feb 25 11:36:09.449: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.38.253 81\nConnection to 10.233.38.253 81 port [tcp/*] succeeded!\n" +Feb 25 11:36:09.449: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" +STEP: Deleting pod pod1 in namespace services-9861 02/25/23 11:36:09.449 +STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-9861 to expose endpoints map[pod2:[101]] 02/25/23 11:36:09.518 +Feb 25 11:36:09.555: INFO: successfully validated that service multi-endpoint-test in namespace services-9861 exposes endpoints map[pod2:[101]] +STEP: Deleting pod pod2 in namespace services-9861 02/25/23 11:36:09.555 +STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-9861 to expose endpoints map[] 02/25/23 11:36:09.615 +Feb 25 11:36:09.662: INFO: successfully validated that service multi-endpoint-test in namespace services-9861 exposes endpoints map[] +[AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 17:06:15.267: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "cronjob-2321" for this suite. 12/14/22 17:06:15.274 -{"msg":"PASSED [sig-apps] CronJob should support CronJob API operations [Conformance]","completed":326,"skipped":6096,"failed":0} +Feb 25 11:36:09.728: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-9861" for this suite. 02/25/23 11:36:09.754 +[AfterEach] [sig-network] Services + test/e2e/network/service.go:762 +{"msg":"PASSED [sig-network] Services should serve multiport endpoints from pods [Conformance]","completed":325,"skipped":6140,"failed":0} ------------------------------ -• [0.218 seconds] -[sig-apps] CronJob -test/e2e/apps/framework.go:23 - should support CronJob API operations [Conformance] - test/e2e/apps/cronjob.go:319 +• [SLOW TEST] [13.201 seconds] +[sig-network] Services +test/e2e/network/common/framework.go:23 + should serve multiport endpoints from pods [Conformance] + test/e2e/network/service.go:852 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] CronJob + [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:06:15.065 - Dec 14 17:06:15.065: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename cronjob 12/14/22 17:06:15.066 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:15.092 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:15.095 - [It] should support CronJob API operations [Conformance] - test/e2e/apps/cronjob.go:319 - STEP: Creating a cronjob 12/14/22 17:06:15.099 - STEP: creating 12/14/22 17:06:15.099 - STEP: getting 12/14/22 17:06:15.111 - STEP: listing 12/14/22 17:06:15.114 - STEP: watching 12/14/22 17:06:15.121 - Dec 14 17:06:15.121: INFO: starting watch - STEP: cluster-wide listing 12/14/22 17:06:15.131 - STEP: cluster-wide watching 12/14/22 17:06:15.137 - Dec 14 17:06:15.137: INFO: starting watch - STEP: patching 12/14/22 17:06:15.138 - STEP: updating 12/14/22 17:06:15.161 - Dec 14 17:06:15.188: INFO: waiting for watch events with expected annotations - Dec 14 17:06:15.189: INFO: saw patched and updated annotations - STEP: patching /status 12/14/22 17:06:15.189 - STEP: updating /status 12/14/22 17:06:15.209 - STEP: get /status 12/14/22 17:06:15.225 - STEP: deleting 12/14/22 17:06:15.228 - STEP: deleting a collection 12/14/22 17:06:15.252 - [AfterEach] [sig-apps] CronJob + STEP: Creating a kubernetes client 02/25/23 11:35:56.568 + Feb 25 11:35:56.568: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 11:35:56.57 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:35:56.677 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:35:56.681 + [BeforeEach] [sig-network] Services + test/e2e/network/service.go:758 + [It] should serve multiport endpoints from pods [Conformance] + test/e2e/network/service.go:852 + STEP: creating service multi-endpoint-test in namespace services-9861 02/25/23 11:35:56.689 + STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-9861 to expose endpoints map[] 02/25/23 11:35:56.749 + Feb 25 11:35:56.773: INFO: successfully validated that service multi-endpoint-test in namespace services-9861 exposes endpoints map[] + STEP: Creating pod pod1 in namespace services-9861 02/25/23 11:35:56.773 + Feb 25 11:35:56.792: INFO: Waiting up to 5m0s for pod "pod1" in namespace "services-9861" to be "running and ready" + Feb 25 11:35:56.822: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 29.691884ms + Feb 25 11:35:56.822: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:35:58.830: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 2.037419176s + Feb 25 11:35:58.830: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:36:00.831: INFO: Pod "pod1": Phase="Running", Reason="", readiness=true. Elapsed: 4.038115024s + Feb 25 11:36:00.831: INFO: The phase of Pod pod1 is Running (Ready = true) + Feb 25 11:36:00.831: INFO: Pod "pod1" satisfied condition "running and ready" + STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-9861 to expose endpoints map[pod1:[100]] 02/25/23 11:36:00.836 + Feb 25 11:36:00.851: INFO: successfully validated that service multi-endpoint-test in namespace services-9861 exposes endpoints map[pod1:[100]] + STEP: Creating pod pod2 in namespace services-9861 02/25/23 11:36:00.852 + Feb 25 11:36:00.921: INFO: Waiting up to 5m0s for pod "pod2" in namespace "services-9861" to be "running and ready" + Feb 25 11:36:00.955: INFO: Pod "pod2": Phase="Pending", Reason="", readiness=false. Elapsed: 34.080947ms + Feb 25 11:36:00.955: INFO: The phase of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:36:02.963: INFO: Pod "pod2": Phase="Running", Reason="", readiness=true. Elapsed: 2.042312421s + Feb 25 11:36:02.964: INFO: The phase of Pod pod2 is Running (Ready = true) + Feb 25 11:36:02.964: INFO: Pod "pod2" satisfied condition "running and ready" + STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-9861 to expose endpoints map[pod1:[100] pod2:[101]] 02/25/23 11:36:02.977 + Feb 25 11:36:03.010: INFO: successfully validated that service multi-endpoint-test in namespace services-9861 exposes endpoints map[pod1:[100] pod2:[101]] + STEP: Checking if the Service forwards traffic to pods 02/25/23 11:36:03.01 + Feb 25 11:36:03.011: INFO: Creating new exec pod + Feb 25 11:36:03.027: INFO: Waiting up to 5m0s for pod "execpod2k2h4" in namespace "services-9861" to be "running" + Feb 25 11:36:03.038: INFO: Pod "execpod2k2h4": Phase="Pending", Reason="", readiness=false. Elapsed: 11.313781ms + Feb 25 11:36:05.049: INFO: Pod "execpod2k2h4": Phase="Pending", Reason="", readiness=false. Elapsed: 2.02216235s + Feb 25 11:36:07.046: INFO: Pod "execpod2k2h4": Phase="Running", Reason="", readiness=true. Elapsed: 4.01866938s + Feb 25 11:36:07.046: INFO: Pod "execpod2k2h4" satisfied condition "running" + Feb 25 11:36:08.047: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9861 exec execpod2k2h4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 multi-endpoint-test 80' + Feb 25 11:36:08.411: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 multi-endpoint-test 80\nConnection to multi-endpoint-test 80 port [tcp/http] succeeded!\n" + Feb 25 11:36:08.411: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:36:08.411: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9861 exec execpod2k2h4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.38.253 80' + Feb 25 11:36:08.733: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.38.253 80\nConnection to 10.233.38.253 80 port [tcp/http] succeeded!\n" + Feb 25 11:36:08.733: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:36:08.733: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9861 exec execpod2k2h4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 multi-endpoint-test 81' + Feb 25 11:36:09.165: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 multi-endpoint-test 81\nConnection to multi-endpoint-test 81 port [tcp/*] succeeded!\n" + Feb 25 11:36:09.165: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + Feb 25 11:36:09.165: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-9861 exec execpod2k2h4 -- /bin/sh -x -c echo hostName | nc -v -t -w 2 10.233.38.253 81' + Feb 25 11:36:09.449: INFO: stderr: "+ echo hostName\n+ nc -v -t -w 2 10.233.38.253 81\nConnection to 10.233.38.253 81 port [tcp/*] succeeded!\n" + Feb 25 11:36:09.449: INFO: stdout: "HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nConnection: close\r\n\r\n400 Bad Request" + STEP: Deleting pod pod1 in namespace services-9861 02/25/23 11:36:09.449 + STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-9861 to expose endpoints map[pod2:[101]] 02/25/23 11:36:09.518 + Feb 25 11:36:09.555: INFO: successfully validated that service multi-endpoint-test in namespace services-9861 exposes endpoints map[pod2:[101]] + STEP: Deleting pod pod2 in namespace services-9861 02/25/23 11:36:09.555 + STEP: waiting up to 3m0s for service multi-endpoint-test in namespace services-9861 to expose endpoints map[] 02/25/23 11:36:09.615 + Feb 25 11:36:09.662: INFO: successfully validated that service multi-endpoint-test in namespace services-9861 exposes endpoints map[] + [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 17:06:15.267: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "cronjob-2321" for this suite. 12/14/22 17:06:15.274 + Feb 25 11:36:09.728: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-9861" for this suite. 02/25/23 11:36:09.754 + [AfterEach] [sig-network] Services + test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSS +SS ------------------------------ -[sig-cli] Kubectl client Kubectl expose - should create services for rc [Conformance] - test/e2e/kubectl/kubectl.go:1413 -[BeforeEach] [sig-cli] Kubectl client +[sig-apps] ReplicaSet + Replace and Patch tests [Conformance] + test/e2e/apps/replica_set.go:154 +[BeforeEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:06:15.285 -Dec 14 17:06:15.285: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 17:06:15.288 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:15.311 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:15.316 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 -[It] should create services for rc [Conformance] - test/e2e/kubectl/kubectl.go:1413 -STEP: creating Agnhost RC 12/14/22 17:06:15.32 -Dec 14 17:06:15.321: INFO: namespace kubectl-3476 -Dec 14 17:06:15.321: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3476 create -f -' -Dec 14 17:06:15.904: INFO: stderr: "" -Dec 14 17:06:15.904: INFO: stdout: "replicationcontroller/agnhost-primary created\n" -STEP: Waiting for Agnhost primary to start. 12/14/22 17:06:15.904 -Dec 14 17:06:16.913: INFO: Selector matched 1 pods for map[app:agnhost] -Dec 14 17:06:16.913: INFO: Found 0 / 1 -Dec 14 17:06:17.912: INFO: Selector matched 1 pods for map[app:agnhost] -Dec 14 17:06:17.912: INFO: Found 0 / 1 -Dec 14 17:06:18.913: INFO: Selector matched 1 pods for map[app:agnhost] -Dec 14 17:06:18.913: INFO: Found 1 / 1 -Dec 14 17:06:18.913: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 -Dec 14 17:06:18.920: INFO: Selector matched 1 pods for map[app:agnhost] -Dec 14 17:06:18.920: INFO: ForEach: Found 1 pods from the filter. Now looping through them. -Dec 14 17:06:18.920: INFO: wait on agnhost-primary startup in kubectl-3476 -Dec 14 17:06:18.920: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3476 logs agnhost-primary-prpz2 agnhost-primary' -Dec 14 17:06:19.118: INFO: stderr: "" -Dec 14 17:06:19.118: INFO: stdout: "Paused\n" -STEP: exposing RC 12/14/22 17:06:19.118 -Dec 14 17:06:19.118: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3476 expose rc agnhost-primary --name=rm2 --port=1234 --target-port=6379' -Dec 14 17:06:19.333: INFO: stderr: "" -Dec 14 17:06:19.333: INFO: stdout: "service/rm2 exposed\n" -Dec 14 17:06:19.340: INFO: Service rm2 in namespace kubectl-3476 found. -STEP: exposing service 12/14/22 17:06:21.353 -Dec 14 17:06:21.354: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3476 expose service rm2 --name=rm3 --port=2345 --target-port=6379' -Dec 14 17:06:21.502: INFO: stderr: "" -Dec 14 17:06:21.502: INFO: stdout: "service/rm3 exposed\n" -Dec 14 17:06:21.532: INFO: Service rm3 in namespace kubectl-3476 found. -[AfterEach] [sig-cli] Kubectl client +STEP: Creating a kubernetes client 02/25/23 11:36:09.774 +Feb 25 11:36:09.775: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename replicaset 02/25/23 11:36:09.781 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:36:09.829 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:36:09.834 +[It] Replace and Patch tests [Conformance] + test/e2e/apps/replica_set.go:154 +Feb 25 11:36:09.891: INFO: Pod name sample-pod: Found 0 pods out of 1 +Feb 25 11:36:14.899: INFO: Pod name sample-pod: Found 1 pods out of 1 +STEP: ensuring each pod is running 02/25/23 11:36:14.899 +STEP: Scaling up "test-rs" replicaset 02/25/23 11:36:14.9 +Feb 25 11:36:14.933: INFO: Updating replica set "test-rs" +STEP: patching the ReplicaSet 02/25/23 11:36:14.933 +W0225 11:36:14.960561 13 warnings.go:70] unknown field "spec.template.spec.TerminationGracePeriodSeconds" +Feb 25 11:36:14.968: INFO: observed ReplicaSet test-rs in namespace replicaset-535 with ReadyReplicas 1, AvailableReplicas 1 +Feb 25 11:36:15.004: INFO: observed ReplicaSet test-rs in namespace replicaset-535 with ReadyReplicas 1, AvailableReplicas 1 +Feb 25 11:36:15.032: INFO: observed ReplicaSet test-rs in namespace replicaset-535 with ReadyReplicas 1, AvailableReplicas 1 +Feb 25 11:36:15.049: INFO: observed ReplicaSet test-rs in namespace replicaset-535 with ReadyReplicas 1, AvailableReplicas 1 +Feb 25 11:36:16.507: INFO: observed ReplicaSet test-rs in namespace replicaset-535 with ReadyReplicas 2, AvailableReplicas 2 +Feb 25 11:36:17.487: INFO: observed Replicaset test-rs in namespace replicaset-535 with ReadyReplicas 3 found true +[AfterEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:187 -Dec 14 17:06:23.551: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-3476" for this suite. 12/14/22 17:06:23.561 -{"msg":"PASSED [sig-cli] Kubectl client Kubectl expose should create services for rc [Conformance]","completed":327,"skipped":6111,"failed":0} +Feb 25 11:36:17.488: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "replicaset-535" for this suite. 02/25/23 11:36:17.495 +{"msg":"PASSED [sig-apps] ReplicaSet Replace and Patch tests [Conformance]","completed":326,"skipped":6142,"failed":0} ------------------------------ -• [SLOW TEST] [8.291 seconds] -[sig-cli] Kubectl client -test/e2e/kubectl/framework.go:23 - Kubectl expose - test/e2e/kubectl/kubectl.go:1407 - should create services for rc [Conformance] - test/e2e/kubectl/kubectl.go:1413 +• [SLOW TEST] [7.736 seconds] +[sig-apps] ReplicaSet +test/e2e/apps/framework.go:23 + Replace and Patch tests [Conformance] + test/e2e/apps/replica_set.go:154 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-cli] Kubectl client + [BeforeEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:06:15.285 - Dec 14 17:06:15.285: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 17:06:15.288 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:15.311 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:15.316 - [BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 - [It] should create services for rc [Conformance] - test/e2e/kubectl/kubectl.go:1413 - STEP: creating Agnhost RC 12/14/22 17:06:15.32 - Dec 14 17:06:15.321: INFO: namespace kubectl-3476 - Dec 14 17:06:15.321: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3476 create -f -' - Dec 14 17:06:15.904: INFO: stderr: "" - Dec 14 17:06:15.904: INFO: stdout: "replicationcontroller/agnhost-primary created\n" - STEP: Waiting for Agnhost primary to start. 12/14/22 17:06:15.904 - Dec 14 17:06:16.913: INFO: Selector matched 1 pods for map[app:agnhost] - Dec 14 17:06:16.913: INFO: Found 0 / 1 - Dec 14 17:06:17.912: INFO: Selector matched 1 pods for map[app:agnhost] - Dec 14 17:06:17.912: INFO: Found 0 / 1 - Dec 14 17:06:18.913: INFO: Selector matched 1 pods for map[app:agnhost] - Dec 14 17:06:18.913: INFO: Found 1 / 1 - Dec 14 17:06:18.913: INFO: WaitFor completed with timeout 5m0s. Pods found = 1 out of 1 - Dec 14 17:06:18.920: INFO: Selector matched 1 pods for map[app:agnhost] - Dec 14 17:06:18.920: INFO: ForEach: Found 1 pods from the filter. Now looping through them. - Dec 14 17:06:18.920: INFO: wait on agnhost-primary startup in kubectl-3476 - Dec 14 17:06:18.920: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3476 logs agnhost-primary-prpz2 agnhost-primary' - Dec 14 17:06:19.118: INFO: stderr: "" - Dec 14 17:06:19.118: INFO: stdout: "Paused\n" - STEP: exposing RC 12/14/22 17:06:19.118 - Dec 14 17:06:19.118: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3476 expose rc agnhost-primary --name=rm2 --port=1234 --target-port=6379' - Dec 14 17:06:19.333: INFO: stderr: "" - Dec 14 17:06:19.333: INFO: stdout: "service/rm2 exposed\n" - Dec 14 17:06:19.340: INFO: Service rm2 in namespace kubectl-3476 found. - STEP: exposing service 12/14/22 17:06:21.353 - Dec 14 17:06:21.354: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3476 expose service rm2 --name=rm3 --port=2345 --target-port=6379' - Dec 14 17:06:21.502: INFO: stderr: "" - Dec 14 17:06:21.502: INFO: stdout: "service/rm3 exposed\n" - Dec 14 17:06:21.532: INFO: Service rm3 in namespace kubectl-3476 found. - [AfterEach] [sig-cli] Kubectl client + STEP: Creating a kubernetes client 02/25/23 11:36:09.774 + Feb 25 11:36:09.775: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename replicaset 02/25/23 11:36:09.781 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:36:09.829 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:36:09.834 + [It] Replace and Patch tests [Conformance] + test/e2e/apps/replica_set.go:154 + Feb 25 11:36:09.891: INFO: Pod name sample-pod: Found 0 pods out of 1 + Feb 25 11:36:14.899: INFO: Pod name sample-pod: Found 1 pods out of 1 + STEP: ensuring each pod is running 02/25/23 11:36:14.899 + STEP: Scaling up "test-rs" replicaset 02/25/23 11:36:14.9 + Feb 25 11:36:14.933: INFO: Updating replica set "test-rs" + STEP: patching the ReplicaSet 02/25/23 11:36:14.933 + W0225 11:36:14.960561 13 warnings.go:70] unknown field "spec.template.spec.TerminationGracePeriodSeconds" + Feb 25 11:36:14.968: INFO: observed ReplicaSet test-rs in namespace replicaset-535 with ReadyReplicas 1, AvailableReplicas 1 + Feb 25 11:36:15.004: INFO: observed ReplicaSet test-rs in namespace replicaset-535 with ReadyReplicas 1, AvailableReplicas 1 + Feb 25 11:36:15.032: INFO: observed ReplicaSet test-rs in namespace replicaset-535 with ReadyReplicas 1, AvailableReplicas 1 + Feb 25 11:36:15.049: INFO: observed ReplicaSet test-rs in namespace replicaset-535 with ReadyReplicas 1, AvailableReplicas 1 + Feb 25 11:36:16.507: INFO: observed ReplicaSet test-rs in namespace replicaset-535 with ReadyReplicas 2, AvailableReplicas 2 + Feb 25 11:36:17.487: INFO: observed Replicaset test-rs in namespace replicaset-535 with ReadyReplicas 3 found true + [AfterEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:187 - Dec 14 17:06:23.551: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-3476" for this suite. 12/14/22 17:06:23.561 + Feb 25 11:36:17.488: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "replicaset-535" for this suite. 02/25/23 11:36:17.495 << End Captured GinkgoWriter Output ------------------------------ -SS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] - custom resource defaulting for requests and from storage works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:269 -[BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +[sig-node] Downward API + should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:216 +[BeforeEach] [sig-node] Downward API test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:06:23.583 -Dec 14 17:06:23.583: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename custom-resource-definition 12/14/22 17:06:23.586 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:23.619 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:23.623 -[It] custom resource defaulting for requests and from storage works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:269 -Dec 14 17:06:23.627: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -[AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 11:36:17.517 +Feb 25 11:36:17.517: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 11:36:17.519 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:36:17.562 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:36:17.567 +[It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:216 +STEP: Creating a pod to test downward api env vars 02/25/23 11:36:17.572 +Feb 25 11:36:17.587: INFO: Waiting up to 5m0s for pod "downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13" in namespace "downward-api-4279" to be "Succeeded or Failed" +Feb 25 11:36:17.594: INFO: Pod "downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13": Phase="Pending", Reason="", readiness=false. Elapsed: 7.271642ms +Feb 25 11:36:19.601: INFO: Pod "downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014812251s +Feb 25 11:36:21.605: INFO: Pod "downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018635461s +STEP: Saw pod success 02/25/23 11:36:21.605 +Feb 25 11:36:21.606: INFO: Pod "downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13" satisfied condition "Succeeded or Failed" +Feb 25 11:36:21.611: INFO: Trying to get logs from node diek3dua9gay-3 pod downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13 container dapi-container: +STEP: delete the pod 02/25/23 11:36:22.014 +Feb 25 11:36:22.033: INFO: Waiting for pod downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13 to disappear +Feb 25 11:36:22.038: INFO: Pod downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13 no longer exists +[AfterEach] [sig-node] Downward API test/e2e/framework/framework.go:187 -Dec 14 17:06:27.084: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "custom-resource-definition-917" for this suite. 12/14/22 17:06:27.093 -{"msg":"PASSED [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] custom resource defaulting for requests and from storage works [Conformance]","completed":328,"skipped":6113,"failed":0} +Feb 25 11:36:22.039: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-4279" for this suite. 02/25/23 11:36:22.046 +{"msg":"PASSED [sig-node] Downward API should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance]","completed":327,"skipped":6183,"failed":0} ------------------------------ -• [3.527 seconds] -[sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - custom resource defaulting for requests and from storage works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:269 +• [4.539 seconds] +[sig-node] Downward API +test/e2e/common/node/framework.go:23 + should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:216 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] + [BeforeEach] [sig-node] Downward API test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:06:23.583 - Dec 14 17:06:23.583: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename custom-resource-definition 12/14/22 17:06:23.586 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:23.619 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:23.623 - [It] custom resource defaulting for requests and from storage works [Conformance] - test/e2e/apimachinery/custom_resource_definition.go:269 - Dec 14 17:06:23.627: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - [AfterEach] [sig-api-machinery] CustomResourceDefinition resources [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 11:36:17.517 + Feb 25 11:36:17.517: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 11:36:17.519 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:36:17.562 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:36:17.567 + [It] should provide default limits.cpu/memory from node allocatable [NodeConformance] [Conformance] + test/e2e/common/node/downwardapi.go:216 + STEP: Creating a pod to test downward api env vars 02/25/23 11:36:17.572 + Feb 25 11:36:17.587: INFO: Waiting up to 5m0s for pod "downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13" in namespace "downward-api-4279" to be "Succeeded or Failed" + Feb 25 11:36:17.594: INFO: Pod "downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13": Phase="Pending", Reason="", readiness=false. Elapsed: 7.271642ms + Feb 25 11:36:19.601: INFO: Pod "downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13": Phase="Pending", Reason="", readiness=false. Elapsed: 2.014812251s + Feb 25 11:36:21.605: INFO: Pod "downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.018635461s + STEP: Saw pod success 02/25/23 11:36:21.605 + Feb 25 11:36:21.606: INFO: Pod "downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13" satisfied condition "Succeeded or Failed" + Feb 25 11:36:21.611: INFO: Trying to get logs from node diek3dua9gay-3 pod downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13 container dapi-container: + STEP: delete the pod 02/25/23 11:36:22.014 + Feb 25 11:36:22.033: INFO: Waiting for pod downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13 to disappear + Feb 25 11:36:22.038: INFO: Pod downward-api-87102d35-64c7-40a3-9c34-45cdfd9e6f13 no longer exists + [AfterEach] [sig-node] Downward API test/e2e/framework/framework.go:187 - Dec 14 17:06:27.084: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "custom-resource-definition-917" for this suite. 12/14/22 17:06:27.093 + Feb 25 11:36:22.039: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-4279" for this suite. 02/25/23 11:36:22.046 << End Captured GinkgoWriter Output ------------------------------ +SSSSSSSSSSSSSSSSSSS +------------------------------ [sig-api-machinery] ResourceQuota - should create a ResourceQuota and capture the life of a replication controller. [Conformance] - test/e2e/apimachinery/resource_quota.go:382 + should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] + test/e2e/apimachinery/resource_quota.go:65 [BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:06:27.111 -Dec 14 17:06:27.111: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename resourcequota 12/14/22 17:06:27.114 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:27.141 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:27.144 -[It] should create a ResourceQuota and capture the life of a replication controller. [Conformance] - test/e2e/apimachinery/resource_quota.go:382 -STEP: Counting existing ResourceQuota 12/14/22 17:06:27.148 -STEP: Creating a ResourceQuota 12/14/22 17:06:32.153 -STEP: Ensuring resource quota status is calculated 12/14/22 17:06:32.164 -STEP: Creating a ReplicationController 12/14/22 17:06:34.169 -STEP: Ensuring resource quota status captures replication controller creation 12/14/22 17:06:34.185 -STEP: Deleting a ReplicationController 12/14/22 17:06:36.19 -STEP: Ensuring resource quota status released usage 12/14/22 17:06:36.213 +STEP: Creating a kubernetes client 02/25/23 11:36:22.058 +Feb 25 11:36:22.058: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename resourcequota 02/25/23 11:36:22.062 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:36:22.09 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:36:22.094 +[It] should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] + test/e2e/apimachinery/resource_quota.go:65 +STEP: Counting existing ResourceQuota 02/25/23 11:36:22.097 +STEP: Creating a ResourceQuota 02/25/23 11:36:27.106 +STEP: Ensuring resource quota status is calculated 02/25/23 11:36:27.115 [AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 -Dec 14 17:06:38.220: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-1855" for this suite. 12/14/22 17:06:38.228 -{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and capture the life of a replication controller. [Conformance]","completed":329,"skipped":6113,"failed":0} +Feb 25 11:36:29.130: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-7693" for this suite. 02/25/23 11:36:29.141 +{"msg":"PASSED [sig-api-machinery] ResourceQuota should create a ResourceQuota and ensure its status is promptly calculated. [Conformance]","completed":328,"skipped":6202,"failed":0} ------------------------------ -• [SLOW TEST] [11.128 seconds] +• [SLOW TEST] [7.111 seconds] [sig-api-machinery] ResourceQuota test/e2e/apimachinery/framework.go:23 - should create a ResourceQuota and capture the life of a replication controller. [Conformance] - test/e2e/apimachinery/resource_quota.go:382 + should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] + test/e2e/apimachinery/resource_quota.go:65 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:06:27.111 - Dec 14 17:06:27.111: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename resourcequota 12/14/22 17:06:27.114 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:27.141 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:27.144 - [It] should create a ResourceQuota and capture the life of a replication controller. [Conformance] - test/e2e/apimachinery/resource_quota.go:382 - STEP: Counting existing ResourceQuota 12/14/22 17:06:27.148 - STEP: Creating a ResourceQuota 12/14/22 17:06:32.153 - STEP: Ensuring resource quota status is calculated 12/14/22 17:06:32.164 - STEP: Creating a ReplicationController 12/14/22 17:06:34.169 - STEP: Ensuring resource quota status captures replication controller creation 12/14/22 17:06:34.185 - STEP: Deleting a ReplicationController 12/14/22 17:06:36.19 - STEP: Ensuring resource quota status released usage 12/14/22 17:06:36.213 + STEP: Creating a kubernetes client 02/25/23 11:36:22.058 + Feb 25 11:36:22.058: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename resourcequota 02/25/23 11:36:22.062 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:36:22.09 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:36:22.094 + [It] should create a ResourceQuota and ensure its status is promptly calculated. [Conformance] + test/e2e/apimachinery/resource_quota.go:65 + STEP: Counting existing ResourceQuota 02/25/23 11:36:22.097 + STEP: Creating a ResourceQuota 02/25/23 11:36:27.106 + STEP: Ensuring resource quota status is calculated 02/25/23 11:36:27.115 [AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 - Dec 14 17:06:38.220: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "resourcequota-1855" for this suite. 12/14/22 17:06:38.228 + Feb 25 11:36:29.130: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "resourcequota-7693" for this suite. 02/25/23 11:36:29.141 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-network] HostPort - validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance] - test/e2e/network/hostport.go:63 -[BeforeEach] [sig-network] HostPort - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:06:38.28 -Dec 14 17:06:38.280: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename hostport 12/14/22 17:06:38.283 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:38.316 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:38.32 -[BeforeEach] [sig-network] HostPort - test/e2e/network/hostport.go:49 -[It] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance] - test/e2e/network/hostport.go:63 -STEP: Trying to create a pod(pod1) with hostport 54323 and hostIP 127.0.0.1 and expect scheduled 12/14/22 17:06:38.329 -Dec 14 17:06:38.339: INFO: Waiting up to 5m0s for pod "pod1" in namespace "hostport-3031" to be "running and ready" -Dec 14 17:06:38.354: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 15.095714ms -Dec 14 17:06:38.355: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 17:06:40.362: INFO: Pod "pod1": Phase="Running", Reason="", readiness=true. Elapsed: 2.022568126s -Dec 14 17:06:40.362: INFO: The phase of Pod pod1 is Running (Ready = true) -Dec 14 17:06:40.362: INFO: Pod "pod1" satisfied condition "running and ready" -STEP: Trying to create another pod(pod2) with hostport 54323 but hostIP 192.168.121.197 on the node which pod1 resides and expect scheduled 12/14/22 17:06:40.362 -Dec 14 17:06:40.385: INFO: Waiting up to 5m0s for pod "pod2" in namespace "hostport-3031" to be "running and ready" -Dec 14 17:06:40.397: INFO: Pod "pod2": Phase="Pending", Reason="", readiness=false. Elapsed: 12.154282ms -Dec 14 17:06:40.397: INFO: The phase of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 17:06:42.406: INFO: Pod "pod2": Phase="Running", Reason="", readiness=true. Elapsed: 2.020943538s -Dec 14 17:06:42.406: INFO: The phase of Pod pod2 is Running (Ready = true) -Dec 14 17:06:42.406: INFO: Pod "pod2" satisfied condition "running and ready" -STEP: Trying to create a third pod(pod3) with hostport 54323, hostIP 192.168.121.197 but use UDP protocol on the node which pod2 resides 12/14/22 17:06:42.406 -Dec 14 17:06:42.414: INFO: Waiting up to 5m0s for pod "pod3" in namespace "hostport-3031" to be "running and ready" -Dec 14 17:06:42.418: INFO: Pod "pod3": Phase="Pending", Reason="", readiness=false. Elapsed: 4.187763ms -Dec 14 17:06:42.418: INFO: The phase of Pod pod3 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 17:06:44.425: INFO: Pod "pod3": Phase="Running", Reason="", readiness=true. Elapsed: 2.011700159s -Dec 14 17:06:44.426: INFO: The phase of Pod pod3 is Running (Ready = true) -Dec 14 17:06:44.426: INFO: Pod "pod3" satisfied condition "running and ready" -Dec 14 17:06:44.434: INFO: Waiting up to 5m0s for pod "e2e-host-exec" in namespace "hostport-3031" to be "running and ready" -Dec 14 17:06:44.438: INFO: Pod "e2e-host-exec": Phase="Pending", Reason="", readiness=false. Elapsed: 4.243446ms -Dec 14 17:06:44.439: INFO: The phase of Pod e2e-host-exec is Pending, waiting for it to be Running (with Ready = true) -Dec 14 17:06:46.448: INFO: Pod "e2e-host-exec": Phase="Running", Reason="", readiness=true. Elapsed: 2.014081172s -Dec 14 17:06:46.448: INFO: The phase of Pod e2e-host-exec is Running (Ready = true) -Dec 14 17:06:46.448: INFO: Pod "e2e-host-exec" satisfied condition "running and ready" -STEP: checking connectivity from pod e2e-host-exec to serverIP: 127.0.0.1, port: 54323 12/14/22 17:06:46.454 -Dec 14 17:06:46.454: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g --connect-timeout 5 --interface 192.168.121.197 http://127.0.0.1:54323/hostname] Namespace:hostport-3031 PodName:e2e-host-exec ContainerName:e2e-host-exec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 17:06:46.454: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 17:06:46.455: INFO: ExecWithOptions: Clientset creation -Dec 14 17:06:46.456: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/hostport-3031/pods/e2e-host-exec/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+--connect-timeout+5+--interface+192.168.121.197+http%3A%2F%2F127.0.0.1%3A54323%2Fhostname&container=e2e-host-exec&container=e2e-host-exec&stderr=true&stdout=true) -STEP: checking connectivity from pod e2e-host-exec to serverIP: 192.168.121.197, port: 54323 12/14/22 17:06:46.583 -Dec 14 17:06:46.583: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g --connect-timeout 5 http://192.168.121.197:54323/hostname] Namespace:hostport-3031 PodName:e2e-host-exec ContainerName:e2e-host-exec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 17:06:46.583: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 17:06:46.585: INFO: ExecWithOptions: Clientset creation -Dec 14 17:06:46.585: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/hostport-3031/pods/e2e-host-exec/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+--connect-timeout+5+http%3A%2F%2F192.168.121.197%3A54323%2Fhostname&container=e2e-host-exec&container=e2e-host-exec&stderr=true&stdout=true) -STEP: checking connectivity from pod e2e-host-exec to serverIP: 192.168.121.197, port: 54323 UDP 12/14/22 17:06:46.678 -Dec 14 17:06:46.678: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostname | nc -u -w 5 192.168.121.197 54323] Namespace:hostport-3031 PodName:e2e-host-exec ContainerName:e2e-host-exec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 17:06:46.678: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 17:06:46.679: INFO: ExecWithOptions: Clientset creation -Dec 14 17:06:46.679: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/hostport-3031/pods/e2e-host-exec/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostname+%7C+nc+-u+-w+5+192.168.121.197+54323&container=e2e-host-exec&container=e2e-host-exec&stderr=true&stdout=true) -[AfterEach] [sig-network] HostPort - test/e2e/framework/framework.go:187 -Dec 14 17:06:51.768: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "hostport-3031" for this suite. 12/14/22 17:06:51.779 -{"msg":"PASSED [sig-network] HostPort validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance]","completed":330,"skipped":6168,"failed":0} ------------------------------- -• [SLOW TEST] [13.513 seconds] -[sig-network] HostPort -test/e2e/network/common/framework.go:23 - validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance] - test/e2e/network/hostport.go:63 - - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] HostPort - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:06:38.28 - Dec 14 17:06:38.280: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename hostport 12/14/22 17:06:38.283 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:38.316 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:38.32 - [BeforeEach] [sig-network] HostPort - test/e2e/network/hostport.go:49 - [It] validates that there is no conflict between pods with same hostPort but different hostIP and protocol [LinuxOnly] [Conformance] - test/e2e/network/hostport.go:63 - STEP: Trying to create a pod(pod1) with hostport 54323 and hostIP 127.0.0.1 and expect scheduled 12/14/22 17:06:38.329 - Dec 14 17:06:38.339: INFO: Waiting up to 5m0s for pod "pod1" in namespace "hostport-3031" to be "running and ready" - Dec 14 17:06:38.354: INFO: Pod "pod1": Phase="Pending", Reason="", readiness=false. Elapsed: 15.095714ms - Dec 14 17:06:38.355: INFO: The phase of Pod pod1 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 17:06:40.362: INFO: Pod "pod1": Phase="Running", Reason="", readiness=true. Elapsed: 2.022568126s - Dec 14 17:06:40.362: INFO: The phase of Pod pod1 is Running (Ready = true) - Dec 14 17:06:40.362: INFO: Pod "pod1" satisfied condition "running and ready" - STEP: Trying to create another pod(pod2) with hostport 54323 but hostIP 192.168.121.197 on the node which pod1 resides and expect scheduled 12/14/22 17:06:40.362 - Dec 14 17:06:40.385: INFO: Waiting up to 5m0s for pod "pod2" in namespace "hostport-3031" to be "running and ready" - Dec 14 17:06:40.397: INFO: Pod "pod2": Phase="Pending", Reason="", readiness=false. Elapsed: 12.154282ms - Dec 14 17:06:40.397: INFO: The phase of Pod pod2 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 17:06:42.406: INFO: Pod "pod2": Phase="Running", Reason="", readiness=true. Elapsed: 2.020943538s - Dec 14 17:06:42.406: INFO: The phase of Pod pod2 is Running (Ready = true) - Dec 14 17:06:42.406: INFO: Pod "pod2" satisfied condition "running and ready" - STEP: Trying to create a third pod(pod3) with hostport 54323, hostIP 192.168.121.197 but use UDP protocol on the node which pod2 resides 12/14/22 17:06:42.406 - Dec 14 17:06:42.414: INFO: Waiting up to 5m0s for pod "pod3" in namespace "hostport-3031" to be "running and ready" - Dec 14 17:06:42.418: INFO: Pod "pod3": Phase="Pending", Reason="", readiness=false. Elapsed: 4.187763ms - Dec 14 17:06:42.418: INFO: The phase of Pod pod3 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 17:06:44.425: INFO: Pod "pod3": Phase="Running", Reason="", readiness=true. Elapsed: 2.011700159s - Dec 14 17:06:44.426: INFO: The phase of Pod pod3 is Running (Ready = true) - Dec 14 17:06:44.426: INFO: Pod "pod3" satisfied condition "running and ready" - Dec 14 17:06:44.434: INFO: Waiting up to 5m0s for pod "e2e-host-exec" in namespace "hostport-3031" to be "running and ready" - Dec 14 17:06:44.438: INFO: Pod "e2e-host-exec": Phase="Pending", Reason="", readiness=false. Elapsed: 4.243446ms - Dec 14 17:06:44.439: INFO: The phase of Pod e2e-host-exec is Pending, waiting for it to be Running (with Ready = true) - Dec 14 17:06:46.448: INFO: Pod "e2e-host-exec": Phase="Running", Reason="", readiness=true. Elapsed: 2.014081172s - Dec 14 17:06:46.448: INFO: The phase of Pod e2e-host-exec is Running (Ready = true) - Dec 14 17:06:46.448: INFO: Pod "e2e-host-exec" satisfied condition "running and ready" - STEP: checking connectivity from pod e2e-host-exec to serverIP: 127.0.0.1, port: 54323 12/14/22 17:06:46.454 - Dec 14 17:06:46.454: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g --connect-timeout 5 --interface 192.168.121.197 http://127.0.0.1:54323/hostname] Namespace:hostport-3031 PodName:e2e-host-exec ContainerName:e2e-host-exec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 17:06:46.454: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 17:06:46.455: INFO: ExecWithOptions: Clientset creation - Dec 14 17:06:46.456: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/hostport-3031/pods/e2e-host-exec/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+--connect-timeout+5+--interface+192.168.121.197+http%3A%2F%2F127.0.0.1%3A54323%2Fhostname&container=e2e-host-exec&container=e2e-host-exec&stderr=true&stdout=true) - STEP: checking connectivity from pod e2e-host-exec to serverIP: 192.168.121.197, port: 54323 12/14/22 17:06:46.583 - Dec 14 17:06:46.583: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g --connect-timeout 5 http://192.168.121.197:54323/hostname] Namespace:hostport-3031 PodName:e2e-host-exec ContainerName:e2e-host-exec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 17:06:46.583: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 17:06:46.585: INFO: ExecWithOptions: Clientset creation - Dec 14 17:06:46.585: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/hostport-3031/pods/e2e-host-exec/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+--connect-timeout+5+http%3A%2F%2F192.168.121.197%3A54323%2Fhostname&container=e2e-host-exec&container=e2e-host-exec&stderr=true&stdout=true) - STEP: checking connectivity from pod e2e-host-exec to serverIP: 192.168.121.197, port: 54323 UDP 12/14/22 17:06:46.678 - Dec 14 17:06:46.678: INFO: ExecWithOptions {Command:[/bin/sh -c echo hostname | nc -u -w 5 192.168.121.197 54323] Namespace:hostport-3031 PodName:e2e-host-exec ContainerName:e2e-host-exec Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 17:06:46.678: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 17:06:46.679: INFO: ExecWithOptions: Clientset creation - Dec 14 17:06:46.679: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/hostport-3031/pods/e2e-host-exec/exec?command=%2Fbin%2Fsh&command=-c&command=echo+hostname+%7C+nc+-u+-w+5+192.168.121.197+54323&container=e2e-host-exec&container=e2e-host-exec&stderr=true&stdout=true) - [AfterEach] [sig-network] HostPort - test/e2e/framework/framework.go:187 - Dec 14 17:06:51.768: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "hostport-3031" for this suite. 12/14/22 17:06:51.779 - << End Captured GinkgoWriter Output +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] ConfigMap - should be immutable if `immutable` field is set [Conformance] - test/e2e/common/storage/configmap_volume.go:503 -[BeforeEach] [sig-storage] ConfigMap +[sig-node] Kubelet when scheduling a read only busybox container + should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:184 +[BeforeEach] [sig-node] Kubelet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:06:51.793 -Dec 14 17:06:51.794: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename configmap 12/14/22 17:06:51.797 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:51.823 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:51.826 -[It] should be immutable if `immutable` field is set [Conformance] - test/e2e/common/storage/configmap_volume.go:503 -[AfterEach] [sig-storage] ConfigMap +STEP: Creating a kubernetes client 02/25/23 11:36:29.185 +Feb 25 11:36:29.185: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubelet-test 02/25/23 11:36:29.187 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:36:29.254 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:36:29.258 +[BeforeEach] [sig-node] Kubelet + test/e2e/common/node/kubelet.go:41 +[It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:184 +Feb 25 11:36:29.289: INFO: Waiting up to 5m0s for pod "busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7" in namespace "kubelet-test-5014" to be "running and ready" +Feb 25 11:36:29.295: INFO: Pod "busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7": Phase="Pending", Reason="", readiness=false. Elapsed: 5.327822ms +Feb 25 11:36:29.295: INFO: The phase of Pod busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:36:31.303: INFO: Pod "busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01415942s +Feb 25 11:36:31.303: INFO: The phase of Pod busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:36:33.302: INFO: Pod "busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7": Phase="Running", Reason="", readiness=true. Elapsed: 4.012641159s +Feb 25 11:36:33.302: INFO: The phase of Pod busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7 is Running (Ready = true) +Feb 25 11:36:33.302: INFO: Pod "busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7" satisfied condition "running and ready" +[AfterEach] [sig-node] Kubelet test/e2e/framework/framework.go:187 -Dec 14 17:06:51.919: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-8438" for this suite. 12/14/22 17:06:51.924 -{"msg":"PASSED [sig-storage] ConfigMap should be immutable if `immutable` field is set [Conformance]","completed":331,"skipped":6168,"failed":0} +Feb 25 11:36:33.319: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubelet-test-5014" for this suite. 02/25/23 11:36:33.327 +{"msg":"PASSED [sig-node] Kubelet when scheduling a read only busybox container should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance]","completed":329,"skipped":6247,"failed":0} ------------------------------ -• [0.145 seconds] -[sig-storage] ConfigMap -test/e2e/common/storage/framework.go:23 - should be immutable if `immutable` field is set [Conformance] - test/e2e/common/storage/configmap_volume.go:503 +• [4.157 seconds] +[sig-node] Kubelet +test/e2e/common/node/framework.go:23 + when scheduling a read only busybox container + test/e2e/common/node/kubelet.go:175 + should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:184 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] ConfigMap + [BeforeEach] [sig-node] Kubelet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:06:51.793 - Dec 14 17:06:51.794: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename configmap 12/14/22 17:06:51.797 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:51.823 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:51.826 - [It] should be immutable if `immutable` field is set [Conformance] - test/e2e/common/storage/configmap_volume.go:503 - [AfterEach] [sig-storage] ConfigMap + STEP: Creating a kubernetes client 02/25/23 11:36:29.185 + Feb 25 11:36:29.185: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubelet-test 02/25/23 11:36:29.187 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:36:29.254 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:36:29.258 + [BeforeEach] [sig-node] Kubelet + test/e2e/common/node/kubelet.go:41 + [It] should not write to root filesystem [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:184 + Feb 25 11:36:29.289: INFO: Waiting up to 5m0s for pod "busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7" in namespace "kubelet-test-5014" to be "running and ready" + Feb 25 11:36:29.295: INFO: Pod "busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7": Phase="Pending", Reason="", readiness=false. Elapsed: 5.327822ms + Feb 25 11:36:29.295: INFO: The phase of Pod busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:36:31.303: INFO: Pod "busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01415942s + Feb 25 11:36:31.303: INFO: The phase of Pod busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:36:33.302: INFO: Pod "busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7": Phase="Running", Reason="", readiness=true. Elapsed: 4.012641159s + Feb 25 11:36:33.302: INFO: The phase of Pod busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7 is Running (Ready = true) + Feb 25 11:36:33.302: INFO: Pod "busybox-readonly-fse564e82a-9b11-4231-aaa5-ecc9739e4fb7" satisfied condition "running and ready" + [AfterEach] [sig-node] Kubelet test/e2e/framework/framework.go:187 - Dec 14 17:06:51.919: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "configmap-8438" for this suite. 12/14/22 17:06:51.924 + Feb 25 11:36:33.319: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubelet-test-5014" for this suite. 02/25/23 11:36:33.327 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSS +SSSSS ------------------------------ -[sig-network] EndpointSliceMirroring - should mirror a custom Endpoints resource through create update and delete [Conformance] - test/e2e/network/endpointslicemirroring.go:53 -[BeforeEach] [sig-network] EndpointSliceMirroring +[sig-storage] Projected downwardAPI + should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:260 +[BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:06:51.941 -Dec 14 17:06:51.941: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename endpointslicemirroring 12/14/22 17:06:51.943 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:51.965 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:51.968 -[BeforeEach] [sig-network] EndpointSliceMirroring - test/e2e/network/endpointslicemirroring.go:41 -[It] should mirror a custom Endpoints resource through create update and delete [Conformance] - test/e2e/network/endpointslicemirroring.go:53 -STEP: mirroring a new custom Endpoint 12/14/22 17:06:51.982 -Dec 14 17:06:51.993: INFO: Waiting for at least 1 EndpointSlice to exist, got 0 -STEP: mirroring an update to a custom Endpoint 12/14/22 17:06:53.998 -Dec 14 17:06:54.008: INFO: Expected EndpointSlice to have 10.2.3.4 as address, got 10.1.2.3 -STEP: mirroring deletion of a custom Endpoint 12/14/22 17:06:56.015 -Dec 14 17:06:56.039: INFO: Waiting for 0 EndpointSlices to exist, got 1 -[AfterEach] [sig-network] EndpointSliceMirroring +STEP: Creating a kubernetes client 02/25/23 11:36:33.346 +Feb 25 11:36:33.346: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 11:36:33.348 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:36:33.397 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:36:33.403 +[BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 +[It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:260 +STEP: Creating a pod to test downward API volume plugin 02/25/23 11:36:33.408 +Feb 25 11:36:33.431: INFO: Waiting up to 5m0s for pod "downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41" in namespace "projected-3877" to be "Succeeded or Failed" +Feb 25 11:36:33.457: INFO: Pod "downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41": Phase="Pending", Reason="", readiness=false. Elapsed: 25.310171ms +Feb 25 11:36:35.464: INFO: Pod "downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033029986s +Feb 25 11:36:37.465: INFO: Pod "downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41": Phase="Pending", Reason="", readiness=false. Elapsed: 4.033303647s +Feb 25 11:36:39.464: INFO: Pod "downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.032374507s +STEP: Saw pod success 02/25/23 11:36:39.464 +Feb 25 11:36:39.464: INFO: Pod "downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41" satisfied condition "Succeeded or Failed" +Feb 25 11:36:39.471: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41 container client-container: +STEP: delete the pod 02/25/23 11:36:39.481 +Feb 25 11:36:39.503: INFO: Waiting for pod downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41 to disappear +Feb 25 11:36:39.521: INFO: Pod downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41 no longer exists +[AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 -Dec 14 17:06:58.054: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "endpointslicemirroring-1316" for this suite. 12/14/22 17:06:58.064 -{"msg":"PASSED [sig-network] EndpointSliceMirroring should mirror a custom Endpoints resource through create update and delete [Conformance]","completed":332,"skipped":6188,"failed":0} +Feb 25 11:36:39.522: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-3877" for this suite. 02/25/23 11:36:39.53 +{"msg":"PASSED [sig-storage] Projected downwardAPI should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance]","completed":330,"skipped":6252,"failed":0} ------------------------------ -• [SLOW TEST] [6.139 seconds] -[sig-network] EndpointSliceMirroring -test/e2e/network/common/framework.go:23 - should mirror a custom Endpoints resource through create update and delete [Conformance] - test/e2e/network/endpointslicemirroring.go:53 +• [SLOW TEST] [6.198 seconds] +[sig-storage] Projected downwardAPI +test/e2e/common/storage/framework.go:23 + should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:260 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] EndpointSliceMirroring + [BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:06:51.941 - Dec 14 17:06:51.941: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename endpointslicemirroring 12/14/22 17:06:51.943 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:51.965 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:51.968 - [BeforeEach] [sig-network] EndpointSliceMirroring - test/e2e/network/endpointslicemirroring.go:41 - [It] should mirror a custom Endpoints resource through create update and delete [Conformance] - test/e2e/network/endpointslicemirroring.go:53 - STEP: mirroring a new custom Endpoint 12/14/22 17:06:51.982 - Dec 14 17:06:51.993: INFO: Waiting for at least 1 EndpointSlice to exist, got 0 - STEP: mirroring an update to a custom Endpoint 12/14/22 17:06:53.998 - Dec 14 17:06:54.008: INFO: Expected EndpointSlice to have 10.2.3.4 as address, got 10.1.2.3 - STEP: mirroring deletion of a custom Endpoint 12/14/22 17:06:56.015 - Dec 14 17:06:56.039: INFO: Waiting for 0 EndpointSlices to exist, got 1 - [AfterEach] [sig-network] EndpointSliceMirroring + STEP: Creating a kubernetes client 02/25/23 11:36:33.346 + Feb 25 11:36:33.346: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 11:36:33.348 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:36:33.397 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:36:33.403 + [BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 + [It] should provide node allocatable (memory) as default memory limit if the limit is not set [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:260 + STEP: Creating a pod to test downward API volume plugin 02/25/23 11:36:33.408 + Feb 25 11:36:33.431: INFO: Waiting up to 5m0s for pod "downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41" in namespace "projected-3877" to be "Succeeded or Failed" + Feb 25 11:36:33.457: INFO: Pod "downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41": Phase="Pending", Reason="", readiness=false. Elapsed: 25.310171ms + Feb 25 11:36:35.464: INFO: Pod "downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033029986s + Feb 25 11:36:37.465: INFO: Pod "downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41": Phase="Pending", Reason="", readiness=false. Elapsed: 4.033303647s + Feb 25 11:36:39.464: INFO: Pod "downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.032374507s + STEP: Saw pod success 02/25/23 11:36:39.464 + Feb 25 11:36:39.464: INFO: Pod "downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41" satisfied condition "Succeeded or Failed" + Feb 25 11:36:39.471: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41 container client-container: + STEP: delete the pod 02/25/23 11:36:39.481 + Feb 25 11:36:39.503: INFO: Waiting for pod downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41 to disappear + Feb 25 11:36:39.521: INFO: Pod downwardapi-volume-920df9f9-3feb-4c00-a798-182416075c41 no longer exists + [AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 - Dec 14 17:06:58.054: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "endpointslicemirroring-1316" for this suite. 12/14/22 17:06:58.064 + Feb 25 11:36:39.522: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-3877" for this suite. 02/25/23 11:36:39.53 << End Captured GinkgoWriter Output ------------------------------ -S +SSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - listing mutating webhooks should work [Conformance] - test/e2e/apimachinery/webhook.go:655 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +[sig-storage] Projected secret + should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:87 +[BeforeEach] [sig-storage] Projected secret test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:06:58.081 -Dec 14 17:06:58.081: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 17:06:58.093 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:58.144 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:58.149 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 17:06:58.175 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 17:06:59.218 -STEP: Deploying the webhook pod 12/14/22 17:06:59.229 -STEP: Wait for the deployment to be ready 12/14/22 17:06:59.244 -Dec 14 17:06:59.252: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set -STEP: Deploying the webhook service 12/14/22 17:07:01.277 -STEP: Verifying the service has paired with the endpoint 12/14/22 17:07:01.295 -Dec 14 17:07:02.295: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] listing mutating webhooks should work [Conformance] - test/e2e/apimachinery/webhook.go:655 -STEP: Listing all of the created validation webhooks 12/14/22 17:07:02.389 -STEP: Creating a configMap that should be mutated 12/14/22 17:07:02.411 -STEP: Deleting the collection of validation webhooks 12/14/22 17:07:02.458 -STEP: Creating a configMap that should not be mutated 12/14/22 17:07:02.545 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 11:36:39.555 +Feb 25 11:36:39.555: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 11:36:39.558 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:36:39.584 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:36:39.588 +[It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:87 +STEP: Creating projection with secret that has name projected-secret-test-map-e9bc62fe-2dcd-4a98-a70f-7faa8ddec015 02/25/23 11:36:39.592 +STEP: Creating a pod to test consume secrets 02/25/23 11:36:39.604 +Feb 25 11:36:39.627: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980" in namespace "projected-6508" to be "Succeeded or Failed" +Feb 25 11:36:39.650: INFO: Pod "pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980": Phase="Pending", Reason="", readiness=false. Elapsed: 22.257607ms +Feb 25 11:36:41.659: INFO: Pod "pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031664442s +Feb 25 11:36:43.659: INFO: Pod "pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980": Phase="Pending", Reason="", readiness=false. Elapsed: 4.031730135s +Feb 25 11:36:45.658: INFO: Pod "pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.029888971s +STEP: Saw pod success 02/25/23 11:36:45.658 +Feb 25 11:36:45.658: INFO: Pod "pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980" satisfied condition "Succeeded or Failed" +Feb 25 11:36:45.664: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980 container projected-secret-volume-test: +STEP: delete the pod 02/25/23 11:36:46.083 +Feb 25 11:36:46.100: INFO: Waiting for pod pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980 to disappear +Feb 25 11:36:46.105: INFO: Pod pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980 no longer exists +[AfterEach] [sig-storage] Projected secret test/e2e/framework/framework.go:187 -Dec 14 17:07:02.561: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-1784" for this suite. 12/14/22 17:07:02.568 -STEP: Destroying namespace "webhook-1784-markers" for this suite. 12/14/22 17:07:02.578 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing mutating webhooks should work [Conformance]","completed":333,"skipped":6189,"failed":0} +Feb 25 11:36:46.106: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-6508" for this suite. 02/25/23 11:36:46.112 +{"msg":"PASSED [sig-storage] Projected secret should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance]","completed":331,"skipped":6278,"failed":0} ------------------------------ -• [4.606 seconds] -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - listing mutating webhooks should work [Conformance] - test/e2e/apimachinery/webhook.go:655 +• [SLOW TEST] [6.566 seconds] +[sig-storage] Projected secret +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:87 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + [BeforeEach] [sig-storage] Projected secret test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:06:58.081 - Dec 14 17:06:58.081: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 17:06:58.093 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:06:58.144 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:06:58.149 - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 17:06:58.175 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 17:06:59.218 - STEP: Deploying the webhook pod 12/14/22 17:06:59.229 - STEP: Wait for the deployment to be ready 12/14/22 17:06:59.244 - Dec 14 17:06:59.252: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set - STEP: Deploying the webhook service 12/14/22 17:07:01.277 - STEP: Verifying the service has paired with the endpoint 12/14/22 17:07:01.295 - Dec 14 17:07:02.295: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] listing mutating webhooks should work [Conformance] - test/e2e/apimachinery/webhook.go:655 - STEP: Listing all of the created validation webhooks 12/14/22 17:07:02.389 - STEP: Creating a configMap that should be mutated 12/14/22 17:07:02.411 - STEP: Deleting the collection of validation webhooks 12/14/22 17:07:02.458 - STEP: Creating a configMap that should not be mutated 12/14/22 17:07:02.545 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 11:36:39.555 + Feb 25 11:36:39.555: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 11:36:39.558 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:36:39.584 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:36:39.588 + [It] should be consumable from pods in volume with mappings and Item Mode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/projected_secret.go:87 + STEP: Creating projection with secret that has name projected-secret-test-map-e9bc62fe-2dcd-4a98-a70f-7faa8ddec015 02/25/23 11:36:39.592 + STEP: Creating a pod to test consume secrets 02/25/23 11:36:39.604 + Feb 25 11:36:39.627: INFO: Waiting up to 5m0s for pod "pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980" in namespace "projected-6508" to be "Succeeded or Failed" + Feb 25 11:36:39.650: INFO: Pod "pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980": Phase="Pending", Reason="", readiness=false. Elapsed: 22.257607ms + Feb 25 11:36:41.659: INFO: Pod "pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031664442s + Feb 25 11:36:43.659: INFO: Pod "pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980": Phase="Pending", Reason="", readiness=false. Elapsed: 4.031730135s + Feb 25 11:36:45.658: INFO: Pod "pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.029888971s + STEP: Saw pod success 02/25/23 11:36:45.658 + Feb 25 11:36:45.658: INFO: Pod "pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980" satisfied condition "Succeeded or Failed" + Feb 25 11:36:45.664: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980 container projected-secret-volume-test: + STEP: delete the pod 02/25/23 11:36:46.083 + Feb 25 11:36:46.100: INFO: Waiting for pod pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980 to disappear + Feb 25 11:36:46.105: INFO: Pod pod-projected-secrets-45e31d48-3cdd-433a-bfa0-977682f50980 no longer exists + [AfterEach] [sig-storage] Projected secret test/e2e/framework/framework.go:187 - Dec 14 17:07:02.561: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-1784" for this suite. 12/14/22 17:07:02.568 - STEP: Destroying namespace "webhook-1784-markers" for this suite. 12/14/22 17:07:02.578 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 + Feb 25 11:36:46.106: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-6508" for this suite. 02/25/23 11:36:46.112 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSS ------------------------------ -[sig-network] DNS - should provide /etc/hosts entries for the cluster [Conformance] - test/e2e/network/dns.go:117 -[BeforeEach] [sig-network] DNS +[sig-storage] EmptyDir wrapper volumes + should not cause race condition when used for configmaps [Serial] [Conformance] + test/e2e/storage/empty_dir_wrapper.go:189 +[BeforeEach] [sig-storage] EmptyDir wrapper volumes test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:07:02.692 -Dec 14 17:07:02.692: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename dns 12/14/22 17:07:02.696 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:02.749 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:02.753 -[It] should provide /etc/hosts entries for the cluster [Conformance] - test/e2e/network/dns.go:117 -STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-1687.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-1687.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;sleep 1; done - 12/14/22 17:07:02.757 -STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-1687.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-1687.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;sleep 1; done - 12/14/22 17:07:02.757 -STEP: creating a pod to probe /etc/hosts 12/14/22 17:07:02.758 -STEP: submitting the pod to kubernetes 12/14/22 17:07:02.758 -Dec 14 17:07:02.783: INFO: Waiting up to 15m0s for pod "dns-test-3f8aec3b-1cac-4b87-bf16-0780b3fd71cd" in namespace "dns-1687" to be "running" -Dec 14 17:07:02.786: INFO: Pod "dns-test-3f8aec3b-1cac-4b87-bf16-0780b3fd71cd": Phase="Pending", Reason="", readiness=false. Elapsed: 3.118562ms -Dec 14 17:07:04.792: INFO: Pod "dns-test-3f8aec3b-1cac-4b87-bf16-0780b3fd71cd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008897377s -Dec 14 17:07:06.799: INFO: Pod "dns-test-3f8aec3b-1cac-4b87-bf16-0780b3fd71cd": Phase="Running", Reason="", readiness=true. Elapsed: 4.016261839s -Dec 14 17:07:06.799: INFO: Pod "dns-test-3f8aec3b-1cac-4b87-bf16-0780b3fd71cd" satisfied condition "running" -STEP: retrieving the pod 12/14/22 17:07:06.8 -STEP: looking for the results for each expected name from probers 12/14/22 17:07:06.81 -Dec 14 17:07:06.838: INFO: DNS probes using dns-1687/dns-test-3f8aec3b-1cac-4b87-bf16-0780b3fd71cd succeeded - -STEP: deleting the pod 12/14/22 17:07:06.838 -[AfterEach] [sig-network] DNS +STEP: Creating a kubernetes client 02/25/23 11:36:46.126 +Feb 25 11:36:46.127: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename emptydir-wrapper 02/25/23 11:36:46.129 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:36:46.156 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:36:46.161 +[It] should not cause race condition when used for configmaps [Serial] [Conformance] + test/e2e/storage/empty_dir_wrapper.go:189 +STEP: Creating 50 configmaps 02/25/23 11:36:46.168 +STEP: Creating RC which spawns configmap-volume pods 02/25/23 11:36:46.555 +Feb 25 11:36:46.577: INFO: Pod name wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d: Found 0 pods out of 5 +Feb 25 11:36:51.595: INFO: Pod name wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d: Found 5 pods out of 5 +STEP: Ensuring each pod is running 02/25/23 11:36:51.595 +Feb 25 11:36:51.595: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh" in namespace "emptydir-wrapper-8741" to be "running" +Feb 25 11:36:51.614: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh": Phase="Pending", Reason="", readiness=false. Elapsed: 18.617595ms +Feb 25 11:36:53.623: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027884865s +Feb 25 11:36:55.623: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh": Phase="Pending", Reason="", readiness=false. Elapsed: 4.027194493s +Feb 25 11:36:57.647: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh": Phase="Pending", Reason="", readiness=false. Elapsed: 6.051922779s +Feb 25 11:36:59.692: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh": Phase="Pending", Reason="", readiness=false. Elapsed: 8.096303846s +Feb 25 11:37:01.621: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh": Phase="Pending", Reason="", readiness=false. Elapsed: 10.02548815s +Feb 25 11:37:03.624: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh": Phase="Running", Reason="", readiness=true. Elapsed: 12.028247813s +Feb 25 11:37:03.624: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh" satisfied condition "running" +Feb 25 11:37:03.624: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-9rpqs" in namespace "emptydir-wrapper-8741" to be "running" +Feb 25 11:37:03.632: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-9rpqs": Phase="Running", Reason="", readiness=true. Elapsed: 7.520421ms +Feb 25 11:37:03.632: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-9rpqs" satisfied condition "running" +Feb 25 11:37:03.632: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-p7tzh" in namespace "emptydir-wrapper-8741" to be "running" +Feb 25 11:37:03.638: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-p7tzh": Phase="Running", Reason="", readiness=true. Elapsed: 6.117926ms +Feb 25 11:37:03.638: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-p7tzh" satisfied condition "running" +Feb 25 11:37:03.638: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-wkkzk" in namespace "emptydir-wrapper-8741" to be "running" +Feb 25 11:37:03.645: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-wkkzk": Phase="Running", Reason="", readiness=true. Elapsed: 6.61725ms +Feb 25 11:37:03.645: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-wkkzk" satisfied condition "running" +Feb 25 11:37:03.645: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-xfsjt" in namespace "emptydir-wrapper-8741" to be "running" +Feb 25 11:37:03.651: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-xfsjt": Phase="Running", Reason="", readiness=true. Elapsed: 5.076496ms +Feb 25 11:37:03.651: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-xfsjt" satisfied condition "running" +STEP: deleting ReplicationController wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d in namespace emptydir-wrapper-8741, will wait for the garbage collector to delete the pods 02/25/23 11:37:03.651 +Feb 25 11:37:03.722: INFO: Deleting ReplicationController wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d took: 14.806022ms +Feb 25 11:37:03.823: INFO: Terminating ReplicationController wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d pods took: 100.969112ms +STEP: Creating RC which spawns configmap-volume pods 02/25/23 11:37:10.432 +Feb 25 11:37:10.452: INFO: Pod name wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb: Found 0 pods out of 5 +Feb 25 11:37:15.472: INFO: Pod name wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb: Found 5 pods out of 5 +STEP: Ensuring each pod is running 02/25/23 11:37:15.472 +Feb 25 11:37:15.472: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp" in namespace "emptydir-wrapper-8741" to be "running" +Feb 25 11:37:15.478: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp": Phase="Pending", Reason="", readiness=false. Elapsed: 6.058085ms +Feb 25 11:37:17.488: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015889085s +Feb 25 11:37:19.486: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp": Phase="Pending", Reason="", readiness=false. Elapsed: 4.014351647s +Feb 25 11:37:21.485: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp": Phase="Pending", Reason="", readiness=false. Elapsed: 6.012720102s +Feb 25 11:37:23.487: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp": Phase="Pending", Reason="", readiness=false. Elapsed: 8.015182887s +Feb 25 11:37:25.488: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp": Phase="Running", Reason="", readiness=true. Elapsed: 10.015725317s +Feb 25 11:37:25.488: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp" satisfied condition "running" +Feb 25 11:37:25.488: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-6ljkg" in namespace "emptydir-wrapper-8741" to be "running" +Feb 25 11:37:25.495: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-6ljkg": Phase="Pending", Reason="", readiness=false. Elapsed: 6.942105ms +Feb 25 11:37:27.503: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-6ljkg": Phase="Running", Reason="", readiness=true. Elapsed: 2.014639257s +Feb 25 11:37:27.503: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-6ljkg" satisfied condition "running" +Feb 25 11:37:27.503: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-ffnpz" in namespace "emptydir-wrapper-8741" to be "running" +Feb 25 11:37:27.508: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-ffnpz": Phase="Running", Reason="", readiness=true. Elapsed: 5.239316ms +Feb 25 11:37:27.508: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-ffnpz" satisfied condition "running" +Feb 25 11:37:27.508: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-g7lrj" in namespace "emptydir-wrapper-8741" to be "running" +Feb 25 11:37:27.515: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-g7lrj": Phase="Running", Reason="", readiness=true. Elapsed: 6.834981ms +Feb 25 11:37:27.515: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-g7lrj" satisfied condition "running" +Feb 25 11:37:27.515: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-pthld" in namespace "emptydir-wrapper-8741" to be "running" +Feb 25 11:37:27.522: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-pthld": Phase="Running", Reason="", readiness=true. Elapsed: 6.795754ms +Feb 25 11:37:27.523: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-pthld" satisfied condition "running" +STEP: deleting ReplicationController wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb in namespace emptydir-wrapper-8741, will wait for the garbage collector to delete the pods 02/25/23 11:37:27.523 +Feb 25 11:37:27.601: INFO: Deleting ReplicationController wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb took: 18.139236ms +Feb 25 11:37:27.702: INFO: Terminating ReplicationController wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb pods took: 100.304744ms +STEP: Creating RC which spawns configmap-volume pods 02/25/23 11:37:33.611 +Feb 25 11:37:33.637: INFO: Pod name wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0: Found 0 pods out of 5 +Feb 25 11:37:38.651: INFO: Pod name wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0: Found 5 pods out of 5 +STEP: Ensuring each pod is running 02/25/23 11:37:38.651 +Feb 25 11:37:38.652: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx" in namespace "emptydir-wrapper-8741" to be "running" +Feb 25 11:37:38.660: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx": Phase="Pending", Reason="", readiness=false. Elapsed: 8.115394ms +Feb 25 11:37:40.669: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01716053s +Feb 25 11:37:42.670: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx": Phase="Pending", Reason="", readiness=false. Elapsed: 4.017845615s +Feb 25 11:37:44.667: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx": Phase="Pending", Reason="", readiness=false. Elapsed: 6.015829978s +Feb 25 11:37:46.672: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx": Phase="Pending", Reason="", readiness=false. Elapsed: 8.019885932s +Feb 25 11:37:48.673: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx": Phase="Pending", Reason="", readiness=false. Elapsed: 10.020962153s +Feb 25 11:37:50.675: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx": Phase="Running", Reason="", readiness=true. Elapsed: 12.023212584s +Feb 25 11:37:50.675: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx" satisfied condition "running" +Feb 25 11:37:50.675: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-jgmdc" in namespace "emptydir-wrapper-8741" to be "running" +Feb 25 11:37:50.683: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-jgmdc": Phase="Running", Reason="", readiness=true. Elapsed: 7.135376ms +Feb 25 11:37:50.683: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-jgmdc" satisfied condition "running" +Feb 25 11:37:50.683: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-pwxkv" in namespace "emptydir-wrapper-8741" to be "running" +Feb 25 11:37:50.688: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-pwxkv": Phase="Running", Reason="", readiness=true. Elapsed: 4.630209ms +Feb 25 11:37:50.688: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-pwxkv" satisfied condition "running" +Feb 25 11:37:50.688: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-s6fqx" in namespace "emptydir-wrapper-8741" to be "running" +Feb 25 11:37:50.695: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-s6fqx": Phase="Running", Reason="", readiness=true. Elapsed: 6.915049ms +Feb 25 11:37:50.695: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-s6fqx" satisfied condition "running" +Feb 25 11:37:50.695: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-wv7bc" in namespace "emptydir-wrapper-8741" to be "running" +Feb 25 11:37:50.702: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-wv7bc": Phase="Running", Reason="", readiness=true. Elapsed: 6.381022ms +Feb 25 11:37:50.702: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-wv7bc" satisfied condition "running" +STEP: deleting ReplicationController wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0 in namespace emptydir-wrapper-8741, will wait for the garbage collector to delete the pods 02/25/23 11:37:50.702 +Feb 25 11:37:50.775: INFO: Deleting ReplicationController wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0 took: 16.741364ms +Feb 25 11:37:50.876: INFO: Terminating ReplicationController wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0 pods took: 101.151951ms +STEP: Cleaning up the configMaps 02/25/23 11:37:55.277 +[AfterEach] [sig-storage] EmptyDir wrapper volumes test/e2e/framework/framework.go:187 -Dec 14 17:07:06.866: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "dns-1687" for this suite. 12/14/22 17:07:06.873 -{"msg":"PASSED [sig-network] DNS should provide /etc/hosts entries for the cluster [Conformance]","completed":334,"skipped":6257,"failed":0} +Feb 25 11:37:55.969: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "emptydir-wrapper-8741" for this suite. 02/25/23 11:37:55.975 +{"msg":"PASSED [sig-storage] EmptyDir wrapper volumes should not cause race condition when used for configmaps [Serial] [Conformance]","completed":332,"skipped":6288,"failed":0} ------------------------------ -• [4.193 seconds] -[sig-network] DNS -test/e2e/network/common/framework.go:23 - should provide /etc/hosts entries for the cluster [Conformance] - test/e2e/network/dns.go:117 +• [SLOW TEST] [69.858 seconds] +[sig-storage] EmptyDir wrapper volumes +test/e2e/storage/utils/framework.go:23 + should not cause race condition when used for configmaps [Serial] [Conformance] + test/e2e/storage/empty_dir_wrapper.go:189 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] DNS + [BeforeEach] [sig-storage] EmptyDir wrapper volumes test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:07:02.692 - Dec 14 17:07:02.692: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename dns 12/14/22 17:07:02.696 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:02.749 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:02.753 - [It] should provide /etc/hosts entries for the cluster [Conformance] - test/e2e/network/dns.go:117 - STEP: Running these commands on wheezy: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-1687.svc.cluster.local)" && echo OK > /results/wheezy_hosts@dns-querier-1.dns-test-service.dns-1687.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/wheezy_hosts@dns-querier-1;sleep 1; done - 12/14/22 17:07:02.757 - STEP: Running these commands on jessie: for i in `seq 1 600`; do test -n "$$(getent hosts dns-querier-1.dns-test-service.dns-1687.svc.cluster.local)" && echo OK > /results/jessie_hosts@dns-querier-1.dns-test-service.dns-1687.svc.cluster.local;test -n "$$(getent hosts dns-querier-1)" && echo OK > /results/jessie_hosts@dns-querier-1;sleep 1; done - 12/14/22 17:07:02.757 - STEP: creating a pod to probe /etc/hosts 12/14/22 17:07:02.758 - STEP: submitting the pod to kubernetes 12/14/22 17:07:02.758 - Dec 14 17:07:02.783: INFO: Waiting up to 15m0s for pod "dns-test-3f8aec3b-1cac-4b87-bf16-0780b3fd71cd" in namespace "dns-1687" to be "running" - Dec 14 17:07:02.786: INFO: Pod "dns-test-3f8aec3b-1cac-4b87-bf16-0780b3fd71cd": Phase="Pending", Reason="", readiness=false. Elapsed: 3.118562ms - Dec 14 17:07:04.792: INFO: Pod "dns-test-3f8aec3b-1cac-4b87-bf16-0780b3fd71cd": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008897377s - Dec 14 17:07:06.799: INFO: Pod "dns-test-3f8aec3b-1cac-4b87-bf16-0780b3fd71cd": Phase="Running", Reason="", readiness=true. Elapsed: 4.016261839s - Dec 14 17:07:06.799: INFO: Pod "dns-test-3f8aec3b-1cac-4b87-bf16-0780b3fd71cd" satisfied condition "running" - STEP: retrieving the pod 12/14/22 17:07:06.8 - STEP: looking for the results for each expected name from probers 12/14/22 17:07:06.81 - Dec 14 17:07:06.838: INFO: DNS probes using dns-1687/dns-test-3f8aec3b-1cac-4b87-bf16-0780b3fd71cd succeeded - - STEP: deleting the pod 12/14/22 17:07:06.838 - [AfterEach] [sig-network] DNS + STEP: Creating a kubernetes client 02/25/23 11:36:46.126 + Feb 25 11:36:46.127: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename emptydir-wrapper 02/25/23 11:36:46.129 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:36:46.156 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:36:46.161 + [It] should not cause race condition when used for configmaps [Serial] [Conformance] + test/e2e/storage/empty_dir_wrapper.go:189 + STEP: Creating 50 configmaps 02/25/23 11:36:46.168 + STEP: Creating RC which spawns configmap-volume pods 02/25/23 11:36:46.555 + Feb 25 11:36:46.577: INFO: Pod name wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d: Found 0 pods out of 5 + Feb 25 11:36:51.595: INFO: Pod name wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d: Found 5 pods out of 5 + STEP: Ensuring each pod is running 02/25/23 11:36:51.595 + Feb 25 11:36:51.595: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh" in namespace "emptydir-wrapper-8741" to be "running" + Feb 25 11:36:51.614: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh": Phase="Pending", Reason="", readiness=false. Elapsed: 18.617595ms + Feb 25 11:36:53.623: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh": Phase="Pending", Reason="", readiness=false. Elapsed: 2.027884865s + Feb 25 11:36:55.623: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh": Phase="Pending", Reason="", readiness=false. Elapsed: 4.027194493s + Feb 25 11:36:57.647: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh": Phase="Pending", Reason="", readiness=false. Elapsed: 6.051922779s + Feb 25 11:36:59.692: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh": Phase="Pending", Reason="", readiness=false. Elapsed: 8.096303846s + Feb 25 11:37:01.621: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh": Phase="Pending", Reason="", readiness=false. Elapsed: 10.02548815s + Feb 25 11:37:03.624: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh": Phase="Running", Reason="", readiness=true. Elapsed: 12.028247813s + Feb 25 11:37:03.624: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-5x7mh" satisfied condition "running" + Feb 25 11:37:03.624: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-9rpqs" in namespace "emptydir-wrapper-8741" to be "running" + Feb 25 11:37:03.632: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-9rpqs": Phase="Running", Reason="", readiness=true. Elapsed: 7.520421ms + Feb 25 11:37:03.632: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-9rpqs" satisfied condition "running" + Feb 25 11:37:03.632: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-p7tzh" in namespace "emptydir-wrapper-8741" to be "running" + Feb 25 11:37:03.638: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-p7tzh": Phase="Running", Reason="", readiness=true. Elapsed: 6.117926ms + Feb 25 11:37:03.638: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-p7tzh" satisfied condition "running" + Feb 25 11:37:03.638: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-wkkzk" in namespace "emptydir-wrapper-8741" to be "running" + Feb 25 11:37:03.645: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-wkkzk": Phase="Running", Reason="", readiness=true. Elapsed: 6.61725ms + Feb 25 11:37:03.645: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-wkkzk" satisfied condition "running" + Feb 25 11:37:03.645: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-xfsjt" in namespace "emptydir-wrapper-8741" to be "running" + Feb 25 11:37:03.651: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-xfsjt": Phase="Running", Reason="", readiness=true. Elapsed: 5.076496ms + Feb 25 11:37:03.651: INFO: Pod "wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d-xfsjt" satisfied condition "running" + STEP: deleting ReplicationController wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d in namespace emptydir-wrapper-8741, will wait for the garbage collector to delete the pods 02/25/23 11:37:03.651 + Feb 25 11:37:03.722: INFO: Deleting ReplicationController wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d took: 14.806022ms + Feb 25 11:37:03.823: INFO: Terminating ReplicationController wrapped-volume-race-6a32375d-719a-4cac-8f63-f86d347a809d pods took: 100.969112ms + STEP: Creating RC which spawns configmap-volume pods 02/25/23 11:37:10.432 + Feb 25 11:37:10.452: INFO: Pod name wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb: Found 0 pods out of 5 + Feb 25 11:37:15.472: INFO: Pod name wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb: Found 5 pods out of 5 + STEP: Ensuring each pod is running 02/25/23 11:37:15.472 + Feb 25 11:37:15.472: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp" in namespace "emptydir-wrapper-8741" to be "running" + Feb 25 11:37:15.478: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp": Phase="Pending", Reason="", readiness=false. Elapsed: 6.058085ms + Feb 25 11:37:17.488: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015889085s + Feb 25 11:37:19.486: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp": Phase="Pending", Reason="", readiness=false. Elapsed: 4.014351647s + Feb 25 11:37:21.485: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp": Phase="Pending", Reason="", readiness=false. Elapsed: 6.012720102s + Feb 25 11:37:23.487: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp": Phase="Pending", Reason="", readiness=false. Elapsed: 8.015182887s + Feb 25 11:37:25.488: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp": Phase="Running", Reason="", readiness=true. Elapsed: 10.015725317s + Feb 25 11:37:25.488: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-59xbp" satisfied condition "running" + Feb 25 11:37:25.488: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-6ljkg" in namespace "emptydir-wrapper-8741" to be "running" + Feb 25 11:37:25.495: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-6ljkg": Phase="Pending", Reason="", readiness=false. Elapsed: 6.942105ms + Feb 25 11:37:27.503: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-6ljkg": Phase="Running", Reason="", readiness=true. Elapsed: 2.014639257s + Feb 25 11:37:27.503: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-6ljkg" satisfied condition "running" + Feb 25 11:37:27.503: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-ffnpz" in namespace "emptydir-wrapper-8741" to be "running" + Feb 25 11:37:27.508: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-ffnpz": Phase="Running", Reason="", readiness=true. Elapsed: 5.239316ms + Feb 25 11:37:27.508: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-ffnpz" satisfied condition "running" + Feb 25 11:37:27.508: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-g7lrj" in namespace "emptydir-wrapper-8741" to be "running" + Feb 25 11:37:27.515: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-g7lrj": Phase="Running", Reason="", readiness=true. Elapsed: 6.834981ms + Feb 25 11:37:27.515: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-g7lrj" satisfied condition "running" + Feb 25 11:37:27.515: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-pthld" in namespace "emptydir-wrapper-8741" to be "running" + Feb 25 11:37:27.522: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-pthld": Phase="Running", Reason="", readiness=true. Elapsed: 6.795754ms + Feb 25 11:37:27.523: INFO: Pod "wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb-pthld" satisfied condition "running" + STEP: deleting ReplicationController wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb in namespace emptydir-wrapper-8741, will wait for the garbage collector to delete the pods 02/25/23 11:37:27.523 + Feb 25 11:37:27.601: INFO: Deleting ReplicationController wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb took: 18.139236ms + Feb 25 11:37:27.702: INFO: Terminating ReplicationController wrapped-volume-race-4b99fc6f-8d35-4128-a34c-d6ce0d0c41bb pods took: 100.304744ms + STEP: Creating RC which spawns configmap-volume pods 02/25/23 11:37:33.611 + Feb 25 11:37:33.637: INFO: Pod name wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0: Found 0 pods out of 5 + Feb 25 11:37:38.651: INFO: Pod name wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0: Found 5 pods out of 5 + STEP: Ensuring each pod is running 02/25/23 11:37:38.651 + Feb 25 11:37:38.652: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx" in namespace "emptydir-wrapper-8741" to be "running" + Feb 25 11:37:38.660: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx": Phase="Pending", Reason="", readiness=false. Elapsed: 8.115394ms + Feb 25 11:37:40.669: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01716053s + Feb 25 11:37:42.670: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx": Phase="Pending", Reason="", readiness=false. Elapsed: 4.017845615s + Feb 25 11:37:44.667: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx": Phase="Pending", Reason="", readiness=false. Elapsed: 6.015829978s + Feb 25 11:37:46.672: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx": Phase="Pending", Reason="", readiness=false. Elapsed: 8.019885932s + Feb 25 11:37:48.673: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx": Phase="Pending", Reason="", readiness=false. Elapsed: 10.020962153s + Feb 25 11:37:50.675: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx": Phase="Running", Reason="", readiness=true. Elapsed: 12.023212584s + Feb 25 11:37:50.675: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-2dqkx" satisfied condition "running" + Feb 25 11:37:50.675: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-jgmdc" in namespace "emptydir-wrapper-8741" to be "running" + Feb 25 11:37:50.683: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-jgmdc": Phase="Running", Reason="", readiness=true. Elapsed: 7.135376ms + Feb 25 11:37:50.683: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-jgmdc" satisfied condition "running" + Feb 25 11:37:50.683: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-pwxkv" in namespace "emptydir-wrapper-8741" to be "running" + Feb 25 11:37:50.688: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-pwxkv": Phase="Running", Reason="", readiness=true. Elapsed: 4.630209ms + Feb 25 11:37:50.688: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-pwxkv" satisfied condition "running" + Feb 25 11:37:50.688: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-s6fqx" in namespace "emptydir-wrapper-8741" to be "running" + Feb 25 11:37:50.695: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-s6fqx": Phase="Running", Reason="", readiness=true. Elapsed: 6.915049ms + Feb 25 11:37:50.695: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-s6fqx" satisfied condition "running" + Feb 25 11:37:50.695: INFO: Waiting up to 5m0s for pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-wv7bc" in namespace "emptydir-wrapper-8741" to be "running" + Feb 25 11:37:50.702: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-wv7bc": Phase="Running", Reason="", readiness=true. Elapsed: 6.381022ms + Feb 25 11:37:50.702: INFO: Pod "wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0-wv7bc" satisfied condition "running" + STEP: deleting ReplicationController wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0 in namespace emptydir-wrapper-8741, will wait for the garbage collector to delete the pods 02/25/23 11:37:50.702 + Feb 25 11:37:50.775: INFO: Deleting ReplicationController wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0 took: 16.741364ms + Feb 25 11:37:50.876: INFO: Terminating ReplicationController wrapped-volume-race-1cf18a16-4a1f-4021-9495-8211e8689ef0 pods took: 101.151951ms + STEP: Cleaning up the configMaps 02/25/23 11:37:55.277 + [AfterEach] [sig-storage] EmptyDir wrapper volumes test/e2e/framework/framework.go:187 - Dec 14 17:07:06.866: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "dns-1687" for this suite. 12/14/22 17:07:06.873 + Feb 25 11:37:55.969: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "emptydir-wrapper-8741" for this suite. 02/25/23 11:37:55.975 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSS +SSSSSS ------------------------------ -[sig-apps] DisruptionController Listing PodDisruptionBudgets for all namespaces - should list and delete a collection of PodDisruptionBudgets [Conformance] - test/e2e/apps/disruption.go:86 -[BeforeEach] [sig-apps] DisruptionController - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:07:06.89 -Dec 14 17:07:06.890: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename disruption 12/14/22 17:07:06.892 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:06.927 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:06.931 -[BeforeEach] [sig-apps] DisruptionController - test/e2e/apps/disruption.go:71 -[BeforeEach] Listing PodDisruptionBudgets for all namespaces +[sig-storage] ConfigMap + should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:56 +[BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:07:06.935 -Dec 14 17:07:06.935: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename disruption-2 12/14/22 17:07:06.937 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:06.963 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:06.968 -[It] should list and delete a collection of PodDisruptionBudgets [Conformance] - test/e2e/apps/disruption.go:86 -STEP: Waiting for the pdb to be processed 12/14/22 17:07:06.981 -STEP: Waiting for the pdb to be processed 12/14/22 17:07:09.005 -STEP: Waiting for the pdb to be processed 12/14/22 17:07:11.026 -STEP: listing a collection of PDBs across all namespaces 12/14/22 17:07:13.037 -STEP: listing a collection of PDBs in namespace disruption-8030 12/14/22 17:07:13.046 -STEP: deleting a collection of PDBs 12/14/22 17:07:13.071 -STEP: Waiting for the PDB collection to be deleted 12/14/22 17:07:13.119 -[AfterEach] Listing PodDisruptionBudgets for all namespaces - test/e2e/framework/framework.go:187 -Dec 14 17:07:13.124: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "disruption-2-8707" for this suite. 12/14/22 17:07:13.131 -[AfterEach] [sig-apps] DisruptionController +STEP: Creating a kubernetes client 02/25/23 11:37:55.986 +Feb 25 11:37:55.987: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename configmap 02/25/23 11:37:55.992 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:37:56.019 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:37:56.022 +[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:56 +STEP: Creating configMap with name configmap-test-volume-2bbfa79a-c418-4adf-8307-1a37ac31c6b3 02/25/23 11:37:56.026 +STEP: Creating a pod to test consume configMaps 02/25/23 11:37:56.033 +Feb 25 11:37:56.044: INFO: Waiting up to 5m0s for pod "pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6" in namespace "configmap-7478" to be "Succeeded or Failed" +Feb 25 11:37:56.049: INFO: Pod "pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6": Phase="Pending", Reason="", readiness=false. Elapsed: 5.58005ms +Feb 25 11:37:58.056: INFO: Pod "pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012615035s +Feb 25 11:38:00.058: INFO: Pod "pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6": Phase="Running", Reason="", readiness=false. Elapsed: 4.014118006s +Feb 25 11:38:02.058: INFO: Pod "pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.013836448s +STEP: Saw pod success 02/25/23 11:38:02.058 +Feb 25 11:38:02.058: INFO: Pod "pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6" satisfied condition "Succeeded or Failed" +Feb 25 11:38:02.065: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6 container agnhost-container: +STEP: delete the pod 02/25/23 11:38:02.118 +Feb 25 11:38:02.142: INFO: Waiting for pod pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6 to disappear +Feb 25 11:38:02.150: INFO: Pod pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6 no longer exists +[AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 -Dec 14 17:07:13.152: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "disruption-8030" for this suite. 12/14/22 17:07:13.16 -{"msg":"PASSED [sig-apps] DisruptionController Listing PodDisruptionBudgets for all namespaces should list and delete a collection of PodDisruptionBudgets [Conformance]","completed":335,"skipped":6270,"failed":0} +Feb 25 11:38:02.151: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-7478" for this suite. 02/25/23 11:38:02.16 +{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","completed":333,"skipped":6294,"failed":0} ------------------------------ -• [SLOW TEST] [6.280 seconds] -[sig-apps] DisruptionController -test/e2e/apps/framework.go:23 - Listing PodDisruptionBudgets for all namespaces - test/e2e/apps/disruption.go:77 - should list and delete a collection of PodDisruptionBudgets [Conformance] - test/e2e/apps/disruption.go:86 +• [SLOW TEST] [6.186 seconds] +[sig-storage] ConfigMap +test/e2e/common/storage/framework.go:23 + should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:56 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] DisruptionController - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:07:06.89 - Dec 14 17:07:06.890: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename disruption 12/14/22 17:07:06.892 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:06.927 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:06.931 - [BeforeEach] [sig-apps] DisruptionController - test/e2e/apps/disruption.go:71 - [BeforeEach] Listing PodDisruptionBudgets for all namespaces + [BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:07:06.935 - Dec 14 17:07:06.935: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename disruption-2 12/14/22 17:07:06.937 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:06.963 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:06.968 - [It] should list and delete a collection of PodDisruptionBudgets [Conformance] - test/e2e/apps/disruption.go:86 - STEP: Waiting for the pdb to be processed 12/14/22 17:07:06.981 - STEP: Waiting for the pdb to be processed 12/14/22 17:07:09.005 - STEP: Waiting for the pdb to be processed 12/14/22 17:07:11.026 - STEP: listing a collection of PDBs across all namespaces 12/14/22 17:07:13.037 - STEP: listing a collection of PDBs in namespace disruption-8030 12/14/22 17:07:13.046 - STEP: deleting a collection of PDBs 12/14/22 17:07:13.071 - STEP: Waiting for the PDB collection to be deleted 12/14/22 17:07:13.119 - [AfterEach] Listing PodDisruptionBudgets for all namespaces - test/e2e/framework/framework.go:187 - Dec 14 17:07:13.124: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "disruption-2-8707" for this suite. 12/14/22 17:07:13.131 - [AfterEach] [sig-apps] DisruptionController + STEP: Creating a kubernetes client 02/25/23 11:37:55.986 + Feb 25 11:37:55.987: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename configmap 02/25/23 11:37:55.992 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:37:56.019 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:37:56.022 + [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:56 + STEP: Creating configMap with name configmap-test-volume-2bbfa79a-c418-4adf-8307-1a37ac31c6b3 02/25/23 11:37:56.026 + STEP: Creating a pod to test consume configMaps 02/25/23 11:37:56.033 + Feb 25 11:37:56.044: INFO: Waiting up to 5m0s for pod "pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6" in namespace "configmap-7478" to be "Succeeded or Failed" + Feb 25 11:37:56.049: INFO: Pod "pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6": Phase="Pending", Reason="", readiness=false. Elapsed: 5.58005ms + Feb 25 11:37:58.056: INFO: Pod "pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012615035s + Feb 25 11:38:00.058: INFO: Pod "pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6": Phase="Running", Reason="", readiness=false. Elapsed: 4.014118006s + Feb 25 11:38:02.058: INFO: Pod "pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.013836448s + STEP: Saw pod success 02/25/23 11:38:02.058 + Feb 25 11:38:02.058: INFO: Pod "pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6" satisfied condition "Succeeded or Failed" + Feb 25 11:38:02.065: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6 container agnhost-container: + STEP: delete the pod 02/25/23 11:38:02.118 + Feb 25 11:38:02.142: INFO: Waiting for pod pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6 to disappear + Feb 25 11:38:02.150: INFO: Pod pod-configmaps-d256ac84-cf66-41b9-b36c-cf0181fc26a6 no longer exists + [AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 - Dec 14 17:07:13.152: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "disruption-8030" for this suite. 12/14/22 17:07:13.16 + Feb 25 11:38:02.151: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "configmap-7478" for this suite. 02/25/23 11:38:02.16 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SS ------------------------------ -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - listing validating webhooks should work [Conformance] - test/e2e/apimachinery/webhook.go:581 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +[sig-node] Security Context When creating a pod with readOnlyRootFilesystem + should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:485 +[BeforeEach] [sig-node] Security Context test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:07:13.175 -Dec 14 17:07:13.176: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 17:07:13.178 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:13.203 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:13.206 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 17:07:13.233 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 17:07:14.139 -STEP: Deploying the webhook pod 12/14/22 17:07:14.146 -STEP: Wait for the deployment to be ready 12/14/22 17:07:14.16 -Dec 14 17:07:14.169: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created -Dec 14 17:07:16.192: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 17, 7, 14, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 7, 14, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 7, 14, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 7, 14, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} -STEP: Deploying the webhook service 12/14/22 17:07:18.198 -STEP: Verifying the service has paired with the endpoint 12/14/22 17:07:18.21 -Dec 14 17:07:19.211: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] listing validating webhooks should work [Conformance] - test/e2e/apimachinery/webhook.go:581 -STEP: Listing all of the created validation webhooks 12/14/22 17:07:19.304 -STEP: Creating a configMap that does not comply to the validation webhook rules 12/14/22 17:07:19.359 -STEP: Deleting the collection of validation webhooks 12/14/22 17:07:19.456 -STEP: Creating a configMap that does not comply to the validation webhook rules 12/14/22 17:07:19.543 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 11:38:02.175 +Feb 25 11:38:02.175: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename security-context-test 02/25/23 11:38:02.178 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:38:02.218 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:38:02.223 +[BeforeEach] [sig-node] Security Context + test/e2e/common/node/security_context.go:49 +[It] should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:485 +Feb 25 11:38:02.242: INFO: Waiting up to 5m0s for pod "busybox-readonly-false-63c738ac-447e-4899-917d-4a0eaf699974" in namespace "security-context-test-6326" to be "Succeeded or Failed" +Feb 25 11:38:02.247: INFO: Pod "busybox-readonly-false-63c738ac-447e-4899-917d-4a0eaf699974": Phase="Pending", Reason="", readiness=false. Elapsed: 4.250509ms +Feb 25 11:38:04.253: INFO: Pod "busybox-readonly-false-63c738ac-447e-4899-917d-4a0eaf699974": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010979763s +Feb 25 11:38:06.254: INFO: Pod "busybox-readonly-false-63c738ac-447e-4899-917d-4a0eaf699974": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011562045s +Feb 25 11:38:08.254: INFO: Pod "busybox-readonly-false-63c738ac-447e-4899-917d-4a0eaf699974": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.011203449s +Feb 25 11:38:08.255: INFO: Pod "busybox-readonly-false-63c738ac-447e-4899-917d-4a0eaf699974" satisfied condition "Succeeded or Failed" +[AfterEach] [sig-node] Security Context test/e2e/framework/framework.go:187 -Dec 14 17:07:19.559: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-5883" for this suite. 12/14/22 17:07:19.566 -STEP: Destroying namespace "webhook-5883-markers" for this suite. 12/14/22 17:07:19.574 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] listing validating webhooks should work [Conformance]","completed":336,"skipped":6303,"failed":0} +Feb 25 11:38:08.257: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "security-context-test-6326" for this suite. 02/25/23 11:38:08.268 +{"msg":"PASSED [sig-node] Security Context When creating a pod with readOnlyRootFilesystem should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance]","completed":334,"skipped":6296,"failed":0} ------------------------------ -• [SLOW TEST] [6.486 seconds] -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - listing validating webhooks should work [Conformance] - test/e2e/apimachinery/webhook.go:581 +• [SLOW TEST] [6.105 seconds] +[sig-node] Security Context +test/e2e/common/node/framework.go:23 + When creating a pod with readOnlyRootFilesystem + test/e2e/common/node/security_context.go:429 + should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:485 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + [BeforeEach] [sig-node] Security Context test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:07:13.175 - Dec 14 17:07:13.176: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 17:07:13.178 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:13.203 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:13.206 - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 17:07:13.233 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 17:07:14.139 - STEP: Deploying the webhook pod 12/14/22 17:07:14.146 - STEP: Wait for the deployment to be ready 12/14/22 17:07:14.16 - Dec 14 17:07:14.169: INFO: new replicaset for deployment "sample-webhook-deployment" is yet to be created - Dec 14 17:07:16.192: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 17, 7, 14, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 7, 14, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 7, 14, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 7, 14, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} - STEP: Deploying the webhook service 12/14/22 17:07:18.198 - STEP: Verifying the service has paired with the endpoint 12/14/22 17:07:18.21 - Dec 14 17:07:19.211: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] listing validating webhooks should work [Conformance] - test/e2e/apimachinery/webhook.go:581 - STEP: Listing all of the created validation webhooks 12/14/22 17:07:19.304 - STEP: Creating a configMap that does not comply to the validation webhook rules 12/14/22 17:07:19.359 - STEP: Deleting the collection of validation webhooks 12/14/22 17:07:19.456 - STEP: Creating a configMap that does not comply to the validation webhook rules 12/14/22 17:07:19.543 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 11:38:02.175 + Feb 25 11:38:02.175: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename security-context-test 02/25/23 11:38:02.178 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:38:02.218 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:38:02.223 + [BeforeEach] [sig-node] Security Context + test/e2e/common/node/security_context.go:49 + [It] should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:485 + Feb 25 11:38:02.242: INFO: Waiting up to 5m0s for pod "busybox-readonly-false-63c738ac-447e-4899-917d-4a0eaf699974" in namespace "security-context-test-6326" to be "Succeeded or Failed" + Feb 25 11:38:02.247: INFO: Pod "busybox-readonly-false-63c738ac-447e-4899-917d-4a0eaf699974": Phase="Pending", Reason="", readiness=false. Elapsed: 4.250509ms + Feb 25 11:38:04.253: INFO: Pod "busybox-readonly-false-63c738ac-447e-4899-917d-4a0eaf699974": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010979763s + Feb 25 11:38:06.254: INFO: Pod "busybox-readonly-false-63c738ac-447e-4899-917d-4a0eaf699974": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011562045s + Feb 25 11:38:08.254: INFO: Pod "busybox-readonly-false-63c738ac-447e-4899-917d-4a0eaf699974": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.011203449s + Feb 25 11:38:08.255: INFO: Pod "busybox-readonly-false-63c738ac-447e-4899-917d-4a0eaf699974" satisfied condition "Succeeded or Failed" + [AfterEach] [sig-node] Security Context test/e2e/framework/framework.go:187 - Dec 14 17:07:19.559: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-5883" for this suite. 12/14/22 17:07:19.566 - STEP: Destroying namespace "webhook-5883-markers" for this suite. 12/14/22 17:07:19.574 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 + Feb 25 11:38:08.257: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "security-context-test-6326" for this suite. 02/25/23 11:38:08.268 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSS +SSSSSSSSSS ------------------------------ -[sig-storage] Projected configMap - should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:56 -[BeforeEach] [sig-storage] Projected configMap +[sig-cli] Kubectl client Kubectl label + should update the label on a resource [Conformance] + test/e2e/kubectl/kubectl.go:1507 +[BeforeEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:07:19.663 -Dec 14 17:07:19.663: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 17:07:19.668 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:19.714 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:19.718 -[It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:56 -STEP: Creating configMap with name projected-configmap-test-volume-9683526b-fcdb-42a4-b48e-854f822240de 12/14/22 17:07:19.721 -STEP: Creating a pod to test consume configMaps 12/14/22 17:07:19.729 -Dec 14 17:07:19.740: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455" in namespace "projected-291" to be "Succeeded or Failed" -Dec 14 17:07:19.754: INFO: Pod "pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455": Phase="Pending", Reason="", readiness=false. Elapsed: 13.720425ms -Dec 14 17:07:21.772: INFO: Pod "pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031813488s -Dec 14 17:07:23.763: INFO: Pod "pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022756331s -STEP: Saw pod success 12/14/22 17:07:23.763 -Dec 14 17:07:23.763: INFO: Pod "pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455" satisfied condition "Succeeded or Failed" -Dec 14 17:07:23.770: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455 container agnhost-container: -STEP: delete the pod 12/14/22 17:07:24.258 -Dec 14 17:07:24.290: INFO: Waiting for pod pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455 to disappear -Dec 14 17:07:24.295: INFO: Pod pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455 no longer exists -[AfterEach] [sig-storage] Projected configMap +STEP: Creating a kubernetes client 02/25/23 11:38:08.281 +Feb 25 11:38:08.282: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubectl 02/25/23 11:38:08.283 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:38:08.31 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:38:08.314 +[BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 +[BeforeEach] Kubectl label + test/e2e/kubectl/kubectl.go:1492 +STEP: creating the pod 02/25/23 11:38:08.319 +Feb 25 11:38:08.320: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 create -f -' +Feb 25 11:38:09.258: INFO: stderr: "" +Feb 25 11:38:09.258: INFO: stdout: "pod/pause created\n" +Feb 25 11:38:09.258: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause] +Feb 25 11:38:09.258: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-5088" to be "running and ready" +Feb 25 11:38:09.265: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 6.759213ms +Feb 25 11:38:09.265: INFO: Error evaluating pod condition running and ready: want pod 'pause' on '' to be 'Running' but was 'Pending' +Feb 25 11:38:11.275: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016465199s +Feb 25 11:38:11.275: INFO: Error evaluating pod condition running and ready: want pod 'pause' on 'diek3dua9gay-3' to be 'Running' but was 'Pending' +Feb 25 11:38:13.282: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 4.023396073s +Feb 25 11:38:13.282: INFO: Pod "pause" satisfied condition "running and ready" +Feb 25 11:38:13.282: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause] +[It] should update the label on a resource [Conformance] + test/e2e/kubectl/kubectl.go:1507 +STEP: adding the label testing-label with value testing-label-value to a pod 02/25/23 11:38:13.282 +Feb 25 11:38:13.282: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 label pods pause testing-label=testing-label-value' +Feb 25 11:38:13.470: INFO: stderr: "" +Feb 25 11:38:13.470: INFO: stdout: "pod/pause labeled\n" +STEP: verifying the pod has the label testing-label with the value testing-label-value 02/25/23 11:38:13.47 +Feb 25 11:38:13.471: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 get pod pause -L testing-label' +Feb 25 11:38:13.659: INFO: stderr: "" +Feb 25 11:38:13.659: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 4s testing-label-value\n" +STEP: removing the label testing-label of a pod 02/25/23 11:38:13.659 +Feb 25 11:38:13.660: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 label pods pause testing-label-' +Feb 25 11:38:13.893: INFO: stderr: "" +Feb 25 11:38:13.893: INFO: stdout: "pod/pause unlabeled\n" +STEP: verifying the pod doesn't have the label testing-label 02/25/23 11:38:13.893 +Feb 25 11:38:13.894: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 get pod pause -L testing-label' +Feb 25 11:38:14.067: INFO: stderr: "" +Feb 25 11:38:14.067: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 5s \n" +[AfterEach] Kubectl label + test/e2e/kubectl/kubectl.go:1498 +STEP: using delete to clean up resources 02/25/23 11:38:14.067 +Feb 25 11:38:14.068: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 delete --grace-period=0 --force -f -' +Feb 25 11:38:14.228: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" +Feb 25 11:38:14.228: INFO: stdout: "pod \"pause\" force deleted\n" +Feb 25 11:38:14.229: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 get rc,svc -l name=pause --no-headers' +Feb 25 11:38:14.426: INFO: stderr: "No resources found in kubectl-5088 namespace.\n" +Feb 25 11:38:14.426: INFO: stdout: "" +Feb 25 11:38:14.429: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 get pods -l name=pause -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' +Feb 25 11:38:14.638: INFO: stderr: "" +Feb 25 11:38:14.638: INFO: stdout: "" +[AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 -Dec 14 17:07:24.295: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-291" for this suite. 12/14/22 17:07:24.302 -{"msg":"PASSED [sig-storage] Projected configMap should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance]","completed":337,"skipped":6323,"failed":0} +Feb 25 11:38:14.638: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubectl-5088" for this suite. 02/25/23 11:38:14.647 +{"msg":"PASSED [sig-cli] Kubectl client Kubectl label should update the label on a resource [Conformance]","completed":335,"skipped":6306,"failed":0} ------------------------------ -• [4.647 seconds] -[sig-storage] Projected configMap -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:56 +• [SLOW TEST] [6.376 seconds] +[sig-cli] Kubectl client +test/e2e/kubectl/framework.go:23 + Kubectl label + test/e2e/kubectl/kubectl.go:1490 + should update the label on a resource [Conformance] + test/e2e/kubectl/kubectl.go:1507 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected configMap - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:07:19.663 - Dec 14 17:07:19.663: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 17:07:19.668 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:19.714 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:19.718 - [It] should be consumable from pods in volume with defaultMode set [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/storage/projected_configmap.go:56 - STEP: Creating configMap with name projected-configmap-test-volume-9683526b-fcdb-42a4-b48e-854f822240de 12/14/22 17:07:19.721 - STEP: Creating a pod to test consume configMaps 12/14/22 17:07:19.729 - Dec 14 17:07:19.740: INFO: Waiting up to 5m0s for pod "pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455" in namespace "projected-291" to be "Succeeded or Failed" - Dec 14 17:07:19.754: INFO: Pod "pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455": Phase="Pending", Reason="", readiness=false. Elapsed: 13.720425ms - Dec 14 17:07:21.772: INFO: Pod "pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455": Phase="Pending", Reason="", readiness=false. Elapsed: 2.031813488s - Dec 14 17:07:23.763: INFO: Pod "pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.022756331s - STEP: Saw pod success 12/14/22 17:07:23.763 - Dec 14 17:07:23.763: INFO: Pod "pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455" satisfied condition "Succeeded or Failed" - Dec 14 17:07:23.770: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455 container agnhost-container: - STEP: delete the pod 12/14/22 17:07:24.258 - Dec 14 17:07:24.290: INFO: Waiting for pod pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455 to disappear - Dec 14 17:07:24.295: INFO: Pod pod-projected-configmaps-a4e03f76-89cf-495e-ac5d-578d48e7d455 no longer exists - [AfterEach] [sig-storage] Projected configMap + [BeforeEach] [sig-cli] Kubectl client + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 11:38:08.281 + Feb 25 11:38:08.282: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubectl 02/25/23 11:38:08.283 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:38:08.31 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:38:08.314 + [BeforeEach] [sig-cli] Kubectl client + test/e2e/kubectl/kubectl.go:272 + [BeforeEach] Kubectl label + test/e2e/kubectl/kubectl.go:1492 + STEP: creating the pod 02/25/23 11:38:08.319 + Feb 25 11:38:08.320: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 create -f -' + Feb 25 11:38:09.258: INFO: stderr: "" + Feb 25 11:38:09.258: INFO: stdout: "pod/pause created\n" + Feb 25 11:38:09.258: INFO: Waiting up to 5m0s for 1 pods to be running and ready: [pause] + Feb 25 11:38:09.258: INFO: Waiting up to 5m0s for pod "pause" in namespace "kubectl-5088" to be "running and ready" + Feb 25 11:38:09.265: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 6.759213ms + Feb 25 11:38:09.265: INFO: Error evaluating pod condition running and ready: want pod 'pause' on '' to be 'Running' but was 'Pending' + Feb 25 11:38:11.275: INFO: Pod "pause": Phase="Pending", Reason="", readiness=false. Elapsed: 2.016465199s + Feb 25 11:38:11.275: INFO: Error evaluating pod condition running and ready: want pod 'pause' on 'diek3dua9gay-3' to be 'Running' but was 'Pending' + Feb 25 11:38:13.282: INFO: Pod "pause": Phase="Running", Reason="", readiness=true. Elapsed: 4.023396073s + Feb 25 11:38:13.282: INFO: Pod "pause" satisfied condition "running and ready" + Feb 25 11:38:13.282: INFO: Wanted all 1 pods to be running and ready. Result: true. Pods: [pause] + [It] should update the label on a resource [Conformance] + test/e2e/kubectl/kubectl.go:1507 + STEP: adding the label testing-label with value testing-label-value to a pod 02/25/23 11:38:13.282 + Feb 25 11:38:13.282: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 label pods pause testing-label=testing-label-value' + Feb 25 11:38:13.470: INFO: stderr: "" + Feb 25 11:38:13.470: INFO: stdout: "pod/pause labeled\n" + STEP: verifying the pod has the label testing-label with the value testing-label-value 02/25/23 11:38:13.47 + Feb 25 11:38:13.471: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 get pod pause -L testing-label' + Feb 25 11:38:13.659: INFO: stderr: "" + Feb 25 11:38:13.659: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 4s testing-label-value\n" + STEP: removing the label testing-label of a pod 02/25/23 11:38:13.659 + Feb 25 11:38:13.660: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 label pods pause testing-label-' + Feb 25 11:38:13.893: INFO: stderr: "" + Feb 25 11:38:13.893: INFO: stdout: "pod/pause unlabeled\n" + STEP: verifying the pod doesn't have the label testing-label 02/25/23 11:38:13.893 + Feb 25 11:38:13.894: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 get pod pause -L testing-label' + Feb 25 11:38:14.067: INFO: stderr: "" + Feb 25 11:38:14.067: INFO: stdout: "NAME READY STATUS RESTARTS AGE TESTING-LABEL\npause 1/1 Running 0 5s \n" + [AfterEach] Kubectl label + test/e2e/kubectl/kubectl.go:1498 + STEP: using delete to clean up resources 02/25/23 11:38:14.067 + Feb 25 11:38:14.068: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 delete --grace-period=0 --force -f -' + Feb 25 11:38:14.228: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" + Feb 25 11:38:14.228: INFO: stdout: "pod \"pause\" force deleted\n" + Feb 25 11:38:14.229: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 get rc,svc -l name=pause --no-headers' + Feb 25 11:38:14.426: INFO: stderr: "No resources found in kubectl-5088 namespace.\n" + Feb 25 11:38:14.426: INFO: stdout: "" + Feb 25 11:38:14.429: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=kubectl-5088 get pods -l name=pause -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' + Feb 25 11:38:14.638: INFO: stderr: "" + Feb 25 11:38:14.638: INFO: stdout: "" + [AfterEach] [sig-cli] Kubectl client test/e2e/framework/framework.go:187 - Dec 14 17:07:24.295: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-291" for this suite. 12/14/22 17:07:24.302 + Feb 25 11:38:14.638: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubectl-5088" for this suite. 02/25/23 11:38:14.647 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] Subpath Atomic writer volumes - should support subpaths with configmap pod with mountPath of existing file [Conformance] - test/e2e/storage/subpath.go:80 -[BeforeEach] [sig-storage] Subpath +[sig-scheduling] SchedulerPreemption [Serial] + validates basic preemption works [Conformance] + test/e2e/scheduling/preemption.go:125 +[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:07:24.315 -Dec 14 17:07:24.315: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename subpath 12/14/22 17:07:24.318 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:24.343 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:24.346 -[BeforeEach] Atomic writer volumes - test/e2e/storage/subpath.go:40 -STEP: Setting up data 12/14/22 17:07:24.35 -[It] should support subpaths with configmap pod with mountPath of existing file [Conformance] - test/e2e/storage/subpath.go:80 -STEP: Creating pod pod-subpath-test-configmap-ss6m 12/14/22 17:07:24.363 -STEP: Creating a pod to test atomic-volume-subpath 12/14/22 17:07:24.363 -Dec 14 17:07:24.375: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-ss6m" in namespace "subpath-5820" to be "Succeeded or Failed" -Dec 14 17:07:24.398: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Pending", Reason="", readiness=false. Elapsed: 23.116456ms -Dec 14 17:07:26.406: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 2.0310096s -Dec 14 17:07:28.409: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 4.034329958s -Dec 14 17:07:30.407: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 6.032703072s -Dec 14 17:07:32.404: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 8.028868326s -Dec 14 17:07:34.404: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 10.028774442s -Dec 14 17:07:36.406: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 12.03080447s -Dec 14 17:07:38.406: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 14.030883053s -Dec 14 17:07:40.407: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 16.032605319s -Dec 14 17:07:42.407: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 18.031804479s -Dec 14 17:07:44.404: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 20.029117441s -Dec 14 17:07:46.404: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=false. Elapsed: 22.02900195s -Dec 14 17:07:48.408: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=false. Elapsed: 24.033387674s -Dec 14 17:07:50.405: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.029804958s -STEP: Saw pod success 12/14/22 17:07:50.405 -Dec 14 17:07:50.405: INFO: Pod "pod-subpath-test-configmap-ss6m" satisfied condition "Succeeded or Failed" -Dec 14 17:07:50.411: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-subpath-test-configmap-ss6m container test-container-subpath-configmap-ss6m: -STEP: delete the pod 12/14/22 17:07:50.424 -Dec 14 17:07:50.446: INFO: Waiting for pod pod-subpath-test-configmap-ss6m to disappear -Dec 14 17:07:50.451: INFO: Pod pod-subpath-test-configmap-ss6m no longer exists -STEP: Deleting pod pod-subpath-test-configmap-ss6m 12/14/22 17:07:50.451 -Dec 14 17:07:50.452: INFO: Deleting pod "pod-subpath-test-configmap-ss6m" in namespace "subpath-5820" -[AfterEach] [sig-storage] Subpath +STEP: Creating a kubernetes client 02/25/23 11:38:14.659 +Feb 25 11:38:14.660: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename sched-preemption 02/25/23 11:38:14.664 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:38:14.689 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:38:14.694 +[BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:92 +Feb 25 11:38:14.721: INFO: Waiting up to 1m0s for all nodes to be ready +Feb 25 11:39:14.798: INFO: Waiting for terminating namespaces to be deleted... +[It] validates basic preemption works [Conformance] + test/e2e/scheduling/preemption.go:125 +STEP: Create pods that use 4/5 of node resources. 02/25/23 11:39:14.802 +Feb 25 11:39:14.834: INFO: Created pod: pod0-0-sched-preemption-low-priority +Feb 25 11:39:14.845: INFO: Created pod: pod0-1-sched-preemption-medium-priority +Feb 25 11:39:14.890: INFO: Created pod: pod1-0-sched-preemption-medium-priority +Feb 25 11:39:14.931: INFO: Created pod: pod1-1-sched-preemption-medium-priority +Feb 25 11:39:14.984: INFO: Created pod: pod2-0-sched-preemption-medium-priority +Feb 25 11:39:14.994: INFO: Created pod: pod2-1-sched-preemption-medium-priority +STEP: Wait for pods to be scheduled. 02/25/23 11:39:14.994 +Feb 25 11:39:14.995: INFO: Waiting up to 5m0s for pod "pod0-0-sched-preemption-low-priority" in namespace "sched-preemption-3554" to be "running" +Feb 25 11:39:15.002: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 7.177711ms +Feb 25 11:39:17.037: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 2.042444766s +Feb 25 11:39:19.014: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 4.01907899s +Feb 25 11:39:21.020: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 6.025234337s +Feb 25 11:39:23.011: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 8.016129223s +Feb 25 11:39:25.049: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 10.054345825s +Feb 25 11:39:27.010: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Running", Reason="", readiness=true. Elapsed: 12.015062032s +Feb 25 11:39:27.010: INFO: Pod "pod0-0-sched-preemption-low-priority" satisfied condition "running" +Feb 25 11:39:27.010: INFO: Waiting up to 5m0s for pod "pod0-1-sched-preemption-medium-priority" in namespace "sched-preemption-3554" to be "running" +Feb 25 11:39:27.015: INFO: Pod "pod0-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 4.321278ms +Feb 25 11:39:27.015: INFO: Pod "pod0-1-sched-preemption-medium-priority" satisfied condition "running" +Feb 25 11:39:27.015: INFO: Waiting up to 5m0s for pod "pod1-0-sched-preemption-medium-priority" in namespace "sched-preemption-3554" to be "running" +Feb 25 11:39:27.018: INFO: Pod "pod1-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 3.5595ms +Feb 25 11:39:27.018: INFO: Pod "pod1-0-sched-preemption-medium-priority" satisfied condition "running" +Feb 25 11:39:27.018: INFO: Waiting up to 5m0s for pod "pod1-1-sched-preemption-medium-priority" in namespace "sched-preemption-3554" to be "running" +Feb 25 11:39:27.023: INFO: Pod "pod1-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 4.454165ms +Feb 25 11:39:27.023: INFO: Pod "pod1-1-sched-preemption-medium-priority" satisfied condition "running" +Feb 25 11:39:27.023: INFO: Waiting up to 5m0s for pod "pod2-0-sched-preemption-medium-priority" in namespace "sched-preemption-3554" to be "running" +Feb 25 11:39:27.027: INFO: Pod "pod2-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 3.967716ms +Feb 25 11:39:27.027: INFO: Pod "pod2-0-sched-preemption-medium-priority" satisfied condition "running" +Feb 25 11:39:27.027: INFO: Waiting up to 5m0s for pod "pod2-1-sched-preemption-medium-priority" in namespace "sched-preemption-3554" to be "running" +Feb 25 11:39:27.030: INFO: Pod "pod2-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 3.024752ms +Feb 25 11:39:27.030: INFO: Pod "pod2-1-sched-preemption-medium-priority" satisfied condition "running" +STEP: Run a high priority pod that has same requirements as that of lower priority pod 02/25/23 11:39:27.03 +Feb 25 11:39:27.039: INFO: Waiting up to 2m0s for pod "preemptor-pod" in namespace "sched-preemption-3554" to be "running" +Feb 25 11:39:27.048: INFO: Pod "preemptor-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 8.660111ms +Feb 25 11:39:29.129: INFO: Pod "preemptor-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.089865089s +Feb 25 11:39:31.053: INFO: Pod "preemptor-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 4.013560393s +Feb 25 11:39:33.054: INFO: Pod "preemptor-pod": Phase="Running", Reason="", readiness=true. Elapsed: 6.015321893s +Feb 25 11:39:33.055: INFO: Pod "preemptor-pod" satisfied condition "running" +[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:187 -Dec 14 17:07:50.458: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "subpath-5820" for this suite. 12/14/22 17:07:50.467 -{"msg":"PASSED [sig-storage] Subpath Atomic writer volumes should support subpaths with configmap pod with mountPath of existing file [Conformance]","completed":338,"skipped":6346,"failed":0} +Feb 25 11:39:33.175: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "sched-preemption-3554" for this suite. 02/25/23 11:39:33.185 +[AfterEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:80 +{"msg":"PASSED [sig-scheduling] SchedulerPreemption [Serial] validates basic preemption works [Conformance]","completed":336,"skipped":6324,"failed":0} ------------------------------ -• [SLOW TEST] [26.161 seconds] -[sig-storage] Subpath -test/e2e/storage/utils/framework.go:23 - Atomic writer volumes - test/e2e/storage/subpath.go:36 - should support subpaths with configmap pod with mountPath of existing file [Conformance] - test/e2e/storage/subpath.go:80 +• [SLOW TEST] [78.691 seconds] +[sig-scheduling] SchedulerPreemption [Serial] +test/e2e/scheduling/framework.go:40 + validates basic preemption works [Conformance] + test/e2e/scheduling/preemption.go:125 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Subpath + [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:07:24.315 - Dec 14 17:07:24.315: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename subpath 12/14/22 17:07:24.318 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:24.343 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:24.346 - [BeforeEach] Atomic writer volumes - test/e2e/storage/subpath.go:40 - STEP: Setting up data 12/14/22 17:07:24.35 - [It] should support subpaths with configmap pod with mountPath of existing file [Conformance] - test/e2e/storage/subpath.go:80 - STEP: Creating pod pod-subpath-test-configmap-ss6m 12/14/22 17:07:24.363 - STEP: Creating a pod to test atomic-volume-subpath 12/14/22 17:07:24.363 - Dec 14 17:07:24.375: INFO: Waiting up to 5m0s for pod "pod-subpath-test-configmap-ss6m" in namespace "subpath-5820" to be "Succeeded or Failed" - Dec 14 17:07:24.398: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Pending", Reason="", readiness=false. Elapsed: 23.116456ms - Dec 14 17:07:26.406: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 2.0310096s - Dec 14 17:07:28.409: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 4.034329958s - Dec 14 17:07:30.407: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 6.032703072s - Dec 14 17:07:32.404: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 8.028868326s - Dec 14 17:07:34.404: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 10.028774442s - Dec 14 17:07:36.406: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 12.03080447s - Dec 14 17:07:38.406: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 14.030883053s - Dec 14 17:07:40.407: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 16.032605319s - Dec 14 17:07:42.407: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 18.031804479s - Dec 14 17:07:44.404: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=true. Elapsed: 20.029117441s - Dec 14 17:07:46.404: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=false. Elapsed: 22.02900195s - Dec 14 17:07:48.408: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Running", Reason="", readiness=false. Elapsed: 24.033387674s - Dec 14 17:07:50.405: INFO: Pod "pod-subpath-test-configmap-ss6m": Phase="Succeeded", Reason="", readiness=false. Elapsed: 26.029804958s - STEP: Saw pod success 12/14/22 17:07:50.405 - Dec 14 17:07:50.405: INFO: Pod "pod-subpath-test-configmap-ss6m" satisfied condition "Succeeded or Failed" - Dec 14 17:07:50.411: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-subpath-test-configmap-ss6m container test-container-subpath-configmap-ss6m: - STEP: delete the pod 12/14/22 17:07:50.424 - Dec 14 17:07:50.446: INFO: Waiting for pod pod-subpath-test-configmap-ss6m to disappear - Dec 14 17:07:50.451: INFO: Pod pod-subpath-test-configmap-ss6m no longer exists - STEP: Deleting pod pod-subpath-test-configmap-ss6m 12/14/22 17:07:50.451 - Dec 14 17:07:50.452: INFO: Deleting pod "pod-subpath-test-configmap-ss6m" in namespace "subpath-5820" - [AfterEach] [sig-storage] Subpath + STEP: Creating a kubernetes client 02/25/23 11:38:14.659 + Feb 25 11:38:14.660: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename sched-preemption 02/25/23 11:38:14.664 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:38:14.689 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:38:14.694 + [BeforeEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:92 + Feb 25 11:38:14.721: INFO: Waiting up to 1m0s for all nodes to be ready + Feb 25 11:39:14.798: INFO: Waiting for terminating namespaces to be deleted... + [It] validates basic preemption works [Conformance] + test/e2e/scheduling/preemption.go:125 + STEP: Create pods that use 4/5 of node resources. 02/25/23 11:39:14.802 + Feb 25 11:39:14.834: INFO: Created pod: pod0-0-sched-preemption-low-priority + Feb 25 11:39:14.845: INFO: Created pod: pod0-1-sched-preemption-medium-priority + Feb 25 11:39:14.890: INFO: Created pod: pod1-0-sched-preemption-medium-priority + Feb 25 11:39:14.931: INFO: Created pod: pod1-1-sched-preemption-medium-priority + Feb 25 11:39:14.984: INFO: Created pod: pod2-0-sched-preemption-medium-priority + Feb 25 11:39:14.994: INFO: Created pod: pod2-1-sched-preemption-medium-priority + STEP: Wait for pods to be scheduled. 02/25/23 11:39:14.994 + Feb 25 11:39:14.995: INFO: Waiting up to 5m0s for pod "pod0-0-sched-preemption-low-priority" in namespace "sched-preemption-3554" to be "running" + Feb 25 11:39:15.002: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 7.177711ms + Feb 25 11:39:17.037: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 2.042444766s + Feb 25 11:39:19.014: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 4.01907899s + Feb 25 11:39:21.020: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 6.025234337s + Feb 25 11:39:23.011: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 8.016129223s + Feb 25 11:39:25.049: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Pending", Reason="", readiness=false. Elapsed: 10.054345825s + Feb 25 11:39:27.010: INFO: Pod "pod0-0-sched-preemption-low-priority": Phase="Running", Reason="", readiness=true. Elapsed: 12.015062032s + Feb 25 11:39:27.010: INFO: Pod "pod0-0-sched-preemption-low-priority" satisfied condition "running" + Feb 25 11:39:27.010: INFO: Waiting up to 5m0s for pod "pod0-1-sched-preemption-medium-priority" in namespace "sched-preemption-3554" to be "running" + Feb 25 11:39:27.015: INFO: Pod "pod0-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 4.321278ms + Feb 25 11:39:27.015: INFO: Pod "pod0-1-sched-preemption-medium-priority" satisfied condition "running" + Feb 25 11:39:27.015: INFO: Waiting up to 5m0s for pod "pod1-0-sched-preemption-medium-priority" in namespace "sched-preemption-3554" to be "running" + Feb 25 11:39:27.018: INFO: Pod "pod1-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 3.5595ms + Feb 25 11:39:27.018: INFO: Pod "pod1-0-sched-preemption-medium-priority" satisfied condition "running" + Feb 25 11:39:27.018: INFO: Waiting up to 5m0s for pod "pod1-1-sched-preemption-medium-priority" in namespace "sched-preemption-3554" to be "running" + Feb 25 11:39:27.023: INFO: Pod "pod1-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 4.454165ms + Feb 25 11:39:27.023: INFO: Pod "pod1-1-sched-preemption-medium-priority" satisfied condition "running" + Feb 25 11:39:27.023: INFO: Waiting up to 5m0s for pod "pod2-0-sched-preemption-medium-priority" in namespace "sched-preemption-3554" to be "running" + Feb 25 11:39:27.027: INFO: Pod "pod2-0-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 3.967716ms + Feb 25 11:39:27.027: INFO: Pod "pod2-0-sched-preemption-medium-priority" satisfied condition "running" + Feb 25 11:39:27.027: INFO: Waiting up to 5m0s for pod "pod2-1-sched-preemption-medium-priority" in namespace "sched-preemption-3554" to be "running" + Feb 25 11:39:27.030: INFO: Pod "pod2-1-sched-preemption-medium-priority": Phase="Running", Reason="", readiness=true. Elapsed: 3.024752ms + Feb 25 11:39:27.030: INFO: Pod "pod2-1-sched-preemption-medium-priority" satisfied condition "running" + STEP: Run a high priority pod that has same requirements as that of lower priority pod 02/25/23 11:39:27.03 + Feb 25 11:39:27.039: INFO: Waiting up to 2m0s for pod "preemptor-pod" in namespace "sched-preemption-3554" to be "running" + Feb 25 11:39:27.048: INFO: Pod "preemptor-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 8.660111ms + Feb 25 11:39:29.129: INFO: Pod "preemptor-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.089865089s + Feb 25 11:39:31.053: INFO: Pod "preemptor-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 4.013560393s + Feb 25 11:39:33.054: INFO: Pod "preemptor-pod": Phase="Running", Reason="", readiness=true. Elapsed: 6.015321893s + Feb 25 11:39:33.055: INFO: Pod "preemptor-pod" satisfied condition "running" + [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] test/e2e/framework/framework.go:187 - Dec 14 17:07:50.458: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "subpath-5820" for this suite. 12/14/22 17:07:50.467 + Feb 25 11:39:33.175: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "sched-preemption-3554" for this suite. 02/25/23 11:39:33.185 + [AfterEach] [sig-scheduling] SchedulerPreemption [Serial] + test/e2e/scheduling/preemption.go:80 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-storage] ConfigMap - binary data should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:174 -[BeforeEach] [sig-storage] ConfigMap +[sig-storage] Projected downwardAPI + should provide podname only [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:52 +[BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:07:50.486 -Dec 14 17:07:50.487: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename configmap 12/14/22 17:07:50.489 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:50.519 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:50.522 -[It] binary data should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:174 -STEP: Creating configMap with name configmap-test-upd-789ce2a5-aa3d-4daf-af40-c89cf2fd2a3e 12/14/22 17:07:50.534 -STEP: Creating the pod 12/14/22 17:07:50.542 -Dec 14 17:07:50.557: INFO: Waiting up to 5m0s for pod "pod-configmaps-e88e85bb-97b1-4f73-b402-58a7fbce9646" in namespace "configmap-2009" to be "running" -Dec 14 17:07:50.563: INFO: Pod "pod-configmaps-e88e85bb-97b1-4f73-b402-58a7fbce9646": Phase="Pending", Reason="", readiness=false. Elapsed: 6.373543ms -Dec 14 17:07:52.571: INFO: Pod "pod-configmaps-e88e85bb-97b1-4f73-b402-58a7fbce9646": Phase="Running", Reason="", readiness=false. Elapsed: 2.014364959s -Dec 14 17:07:52.571: INFO: Pod "pod-configmaps-e88e85bb-97b1-4f73-b402-58a7fbce9646" satisfied condition "running" -STEP: Waiting for pod with text data 12/14/22 17:07:52.571 -STEP: Waiting for pod with binary data 12/14/22 17:07:52.581 -[AfterEach] [sig-storage] ConfigMap +STEP: Creating a kubernetes client 02/25/23 11:39:33.354 +Feb 25 11:39:33.355: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename projected 02/25/23 11:39:33.357 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:39:33.392 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:39:33.399 +[BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 +[It] should provide podname only [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:52 +STEP: Creating a pod to test downward API volume plugin 02/25/23 11:39:33.404 +Feb 25 11:39:33.420: INFO: Waiting up to 5m0s for pod "downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d" in namespace "projected-9459" to be "Succeeded or Failed" +Feb 25 11:39:33.425: INFO: Pod "downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d": Phase="Pending", Reason="", readiness=false. Elapsed: 5.253502ms +Feb 25 11:39:35.436: INFO: Pod "downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015900777s +Feb 25 11:39:37.434: INFO: Pod "downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d": Phase="Pending", Reason="", readiness=false. Elapsed: 4.014789533s +Feb 25 11:39:39.431: INFO: Pod "downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.011031814s +STEP: Saw pod success 02/25/23 11:39:39.431 +Feb 25 11:39:39.431: INFO: Pod "downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d" satisfied condition "Succeeded or Failed" +Feb 25 11:39:39.437: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d container client-container: +STEP: delete the pod 02/25/23 11:39:39.463 +Feb 25 11:39:39.482: INFO: Waiting for pod downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d to disappear +Feb 25 11:39:39.498: INFO: Pod downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d no longer exists +[AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 -Dec 14 17:07:52.593: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-2009" for this suite. 12/14/22 17:07:52.601 -{"msg":"PASSED [sig-storage] ConfigMap binary data should be reflected in volume [NodeConformance] [Conformance]","completed":339,"skipped":6361,"failed":0} +Feb 25 11:39:39.500: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "projected-9459" for this suite. 02/25/23 11:39:39.508 +{"msg":"PASSED [sig-storage] Projected downwardAPI should provide podname only [NodeConformance] [Conformance]","completed":337,"skipped":6370,"failed":0} ------------------------------ -• [2.125 seconds] -[sig-storage] ConfigMap +• [SLOW TEST] [6.171 seconds] +[sig-storage] Projected downwardAPI test/e2e/common/storage/framework.go:23 - binary data should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:174 + should provide podname only [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:52 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] ConfigMap + [BeforeEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:07:50.486 - Dec 14 17:07:50.487: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename configmap 12/14/22 17:07:50.489 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:50.519 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:50.522 - [It] binary data should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:174 - STEP: Creating configMap with name configmap-test-upd-789ce2a5-aa3d-4daf-af40-c89cf2fd2a3e 12/14/22 17:07:50.534 - STEP: Creating the pod 12/14/22 17:07:50.542 - Dec 14 17:07:50.557: INFO: Waiting up to 5m0s for pod "pod-configmaps-e88e85bb-97b1-4f73-b402-58a7fbce9646" in namespace "configmap-2009" to be "running" - Dec 14 17:07:50.563: INFO: Pod "pod-configmaps-e88e85bb-97b1-4f73-b402-58a7fbce9646": Phase="Pending", Reason="", readiness=false. Elapsed: 6.373543ms - Dec 14 17:07:52.571: INFO: Pod "pod-configmaps-e88e85bb-97b1-4f73-b402-58a7fbce9646": Phase="Running", Reason="", readiness=false. Elapsed: 2.014364959s - Dec 14 17:07:52.571: INFO: Pod "pod-configmaps-e88e85bb-97b1-4f73-b402-58a7fbce9646" satisfied condition "running" - STEP: Waiting for pod with text data 12/14/22 17:07:52.571 - STEP: Waiting for pod with binary data 12/14/22 17:07:52.581 - [AfterEach] [sig-storage] ConfigMap + STEP: Creating a kubernetes client 02/25/23 11:39:33.354 + Feb 25 11:39:33.355: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename projected 02/25/23 11:39:33.357 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:39:33.392 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:39:33.399 + [BeforeEach] [sig-storage] Projected downwardAPI + test/e2e/common/storage/projected_downwardapi.go:43 + [It] should provide podname only [NodeConformance] [Conformance] + test/e2e/common/storage/projected_downwardapi.go:52 + STEP: Creating a pod to test downward API volume plugin 02/25/23 11:39:33.404 + Feb 25 11:39:33.420: INFO: Waiting up to 5m0s for pod "downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d" in namespace "projected-9459" to be "Succeeded or Failed" + Feb 25 11:39:33.425: INFO: Pod "downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d": Phase="Pending", Reason="", readiness=false. Elapsed: 5.253502ms + Feb 25 11:39:35.436: INFO: Pod "downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015900777s + Feb 25 11:39:37.434: INFO: Pod "downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d": Phase="Pending", Reason="", readiness=false. Elapsed: 4.014789533s + Feb 25 11:39:39.431: INFO: Pod "downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.011031814s + STEP: Saw pod success 02/25/23 11:39:39.431 + Feb 25 11:39:39.431: INFO: Pod "downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d" satisfied condition "Succeeded or Failed" + Feb 25 11:39:39.437: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d container client-container: + STEP: delete the pod 02/25/23 11:39:39.463 + Feb 25 11:39:39.482: INFO: Waiting for pod downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d to disappear + Feb 25 11:39:39.498: INFO: Pod downwardapi-volume-76616170-fa2e-4ea4-a954-8da09497dd5d no longer exists + [AfterEach] [sig-storage] Projected downwardAPI test/e2e/framework/framework.go:187 - Dec 14 17:07:52.593: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "configmap-2009" for this suite. 12/14/22 17:07:52.601 + Feb 25 11:39:39.500: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "projected-9459" for this suite. 02/25/23 11:39:39.508 << End Captured GinkgoWriter Output ------------------------------ -S +SSS ------------------------------ -[sig-node] Sysctls [LinuxOnly] [NodeConformance] - should support sysctls [MinimumKubeletVersion:1.21] [Conformance] - test/e2e/common/node/sysctl.go:77 -[BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] - test/e2e/common/node/sysctl.go:37 -[BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] +[sig-node] Container Runtime blackbox test when starting a container that exits + should run with the expected status [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:51 +[BeforeEach] [sig-node] Container Runtime test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:07:52.612 -Dec 14 17:07:52.612: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename sysctl 12/14/22 17:07:52.615 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:52.639 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:52.643 -[BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] - test/e2e/common/node/sysctl.go:67 -[It] should support sysctls [MinimumKubeletVersion:1.21] [Conformance] - test/e2e/common/node/sysctl.go:77 -STEP: Creating a pod with the kernel.shm_rmid_forced sysctl 12/14/22 17:07:52.648 -STEP: Watching for error events or started pod 12/14/22 17:07:52.661 -STEP: Waiting for pod completion 12/14/22 17:07:54.669 -Dec 14 17:07:54.669: INFO: Waiting up to 3m0s for pod "sysctl-907bfe2e-acd0-44ba-bca2-41487c59b91b" in namespace "sysctl-9826" to be "completed" -Dec 14 17:07:54.674: INFO: Pod "sysctl-907bfe2e-acd0-44ba-bca2-41487c59b91b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.964786ms -Dec 14 17:07:56.687: INFO: Pod "sysctl-907bfe2e-acd0-44ba-bca2-41487c59b91b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01740745s -Dec 14 17:07:58.680: INFO: Pod "sysctl-907bfe2e-acd0-44ba-bca2-41487c59b91b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01090616s -Dec 14 17:07:58.681: INFO: Pod "sysctl-907bfe2e-acd0-44ba-bca2-41487c59b91b" satisfied condition "completed" -STEP: Checking that the pod succeeded 12/14/22 17:07:58.688 -STEP: Getting logs from the pod 12/14/22 17:07:58.688 -STEP: Checking that the sysctl is actually updated 12/14/22 17:07:58.698 -[AfterEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] +STEP: Creating a kubernetes client 02/25/23 11:39:39.527 +Feb 25 11:39:39.527: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename container-runtime 02/25/23 11:39:39.529 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:39:39.572 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:39:39.577 +[It] should run with the expected status [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:51 +STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount' 02/25/23 11:39:39.622 +STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase' 02/25/23 11:39:58.77 +STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition 02/25/23 11:39:58.775 +STEP: Container 'terminate-cmd-rpa': should get the expected 'State' 02/25/23 11:39:58.787 +STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance] 02/25/23 11:39:58.787 +STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount' 02/25/23 11:39:58.827 +STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase' 02/25/23 11:40:01.854 +STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition 02/25/23 11:40:03.874 +STEP: Container 'terminate-cmd-rpof': should get the expected 'State' 02/25/23 11:40:03.885 +STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance] 02/25/23 11:40:03.885 +STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount' 02/25/23 11:40:03.918 +STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase' 02/25/23 11:40:04.933 +STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition 02/25/23 11:40:09.971 +STEP: Container 'terminate-cmd-rpn': should get the expected 'State' 02/25/23 11:40:09.981 +STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance] 02/25/23 11:40:09.981 +[AfterEach] [sig-node] Container Runtime test/e2e/framework/framework.go:187 -Dec 14 17:07:58.698: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "sysctl-9826" for this suite. 12/14/22 17:07:58.706 -{"msg":"PASSED [sig-node] Sysctls [LinuxOnly] [NodeConformance] should support sysctls [MinimumKubeletVersion:1.21] [Conformance]","completed":340,"skipped":6362,"failed":0} +Feb 25 11:40:10.017: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "container-runtime-8422" for this suite. 02/25/23 11:40:10.024 +{"msg":"PASSED [sig-node] Container Runtime blackbox test when starting a container that exits should run with the expected status [NodeConformance] [Conformance]","completed":338,"skipped":6373,"failed":0} ------------------------------ -• [SLOW TEST] [6.104 seconds] -[sig-node] Sysctls [LinuxOnly] [NodeConformance] +• [SLOW TEST] [30.508 seconds] +[sig-node] Container Runtime test/e2e/common/node/framework.go:23 - should support sysctls [MinimumKubeletVersion:1.21] [Conformance] - test/e2e/common/node/sysctl.go:77 + blackbox test + test/e2e/common/node/runtime.go:43 + when starting a container that exits + test/e2e/common/node/runtime.go:44 + should run with the expected status [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:51 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] - test/e2e/common/node/sysctl.go:37 - [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] + [BeforeEach] [sig-node] Container Runtime test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:07:52.612 - Dec 14 17:07:52.612: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename sysctl 12/14/22 17:07:52.615 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:52.639 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:52.643 - [BeforeEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] - test/e2e/common/node/sysctl.go:67 - [It] should support sysctls [MinimumKubeletVersion:1.21] [Conformance] - test/e2e/common/node/sysctl.go:77 - STEP: Creating a pod with the kernel.shm_rmid_forced sysctl 12/14/22 17:07:52.648 - STEP: Watching for error events or started pod 12/14/22 17:07:52.661 - STEP: Waiting for pod completion 12/14/22 17:07:54.669 - Dec 14 17:07:54.669: INFO: Waiting up to 3m0s for pod "sysctl-907bfe2e-acd0-44ba-bca2-41487c59b91b" in namespace "sysctl-9826" to be "completed" - Dec 14 17:07:54.674: INFO: Pod "sysctl-907bfe2e-acd0-44ba-bca2-41487c59b91b": Phase="Pending", Reason="", readiness=false. Elapsed: 4.964786ms - Dec 14 17:07:56.687: INFO: Pod "sysctl-907bfe2e-acd0-44ba-bca2-41487c59b91b": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01740745s - Dec 14 17:07:58.680: INFO: Pod "sysctl-907bfe2e-acd0-44ba-bca2-41487c59b91b": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.01090616s - Dec 14 17:07:58.681: INFO: Pod "sysctl-907bfe2e-acd0-44ba-bca2-41487c59b91b" satisfied condition "completed" - STEP: Checking that the pod succeeded 12/14/22 17:07:58.688 - STEP: Getting logs from the pod 12/14/22 17:07:58.688 - STEP: Checking that the sysctl is actually updated 12/14/22 17:07:58.698 - [AfterEach] [sig-node] Sysctls [LinuxOnly] [NodeConformance] + STEP: Creating a kubernetes client 02/25/23 11:39:39.527 + Feb 25 11:39:39.527: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename container-runtime 02/25/23 11:39:39.529 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:39:39.572 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:39:39.577 + [It] should run with the expected status [NodeConformance] [Conformance] + test/e2e/common/node/runtime.go:51 + STEP: Container 'terminate-cmd-rpa': should get the expected 'RestartCount' 02/25/23 11:39:39.622 + STEP: Container 'terminate-cmd-rpa': should get the expected 'Phase' 02/25/23 11:39:58.77 + STEP: Container 'terminate-cmd-rpa': should get the expected 'Ready' condition 02/25/23 11:39:58.775 + STEP: Container 'terminate-cmd-rpa': should get the expected 'State' 02/25/23 11:39:58.787 + STEP: Container 'terminate-cmd-rpa': should be possible to delete [NodeConformance] 02/25/23 11:39:58.787 + STEP: Container 'terminate-cmd-rpof': should get the expected 'RestartCount' 02/25/23 11:39:58.827 + STEP: Container 'terminate-cmd-rpof': should get the expected 'Phase' 02/25/23 11:40:01.854 + STEP: Container 'terminate-cmd-rpof': should get the expected 'Ready' condition 02/25/23 11:40:03.874 + STEP: Container 'terminate-cmd-rpof': should get the expected 'State' 02/25/23 11:40:03.885 + STEP: Container 'terminate-cmd-rpof': should be possible to delete [NodeConformance] 02/25/23 11:40:03.885 + STEP: Container 'terminate-cmd-rpn': should get the expected 'RestartCount' 02/25/23 11:40:03.918 + STEP: Container 'terminate-cmd-rpn': should get the expected 'Phase' 02/25/23 11:40:04.933 + STEP: Container 'terminate-cmd-rpn': should get the expected 'Ready' condition 02/25/23 11:40:09.971 + STEP: Container 'terminate-cmd-rpn': should get the expected 'State' 02/25/23 11:40:09.981 + STEP: Container 'terminate-cmd-rpn': should be possible to delete [NodeConformance] 02/25/23 11:40:09.981 + [AfterEach] [sig-node] Container Runtime test/e2e/framework/framework.go:187 - Dec 14 17:07:58.698: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "sysctl-9826" for this suite. 12/14/22 17:07:58.706 + Feb 25 11:40:10.017: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "container-runtime-8422" for this suite. 02/25/23 11:40:10.024 << End Captured GinkgoWriter Output ------------------------------ -SSSS +SSSSSSSS ------------------------------ -[sig-network] Networking Granular Checks: Pods - should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:105 -[BeforeEach] [sig-network] Networking +[sig-node] Kubelet when scheduling a busybox command that always fails in a pod + should have an terminated reason [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:110 +[BeforeEach] [sig-node] Kubelet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:07:58.717 -Dec 14 17:07:58.717: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename pod-network-test 12/14/22 17:07:58.721 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:58.748 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:58.752 -[It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:105 -STEP: Performing setup for networking test in namespace pod-network-test-3360 12/14/22 17:07:58.755 -STEP: creating a selector 12/14/22 17:07:58.756 -STEP: Creating the service pods in kubernetes 12/14/22 17:07:58.756 -Dec 14 17:07:58.756: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable -Dec 14 17:07:58.792: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-3360" to be "running and ready" -Dec 14 17:07:58.795: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 3.469322ms -Dec 14 17:07:58.795: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 17:08:00.807: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015012934s -Dec 14 17:08:00.807: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 17:08:02.803: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.011744195s -Dec 14 17:08:02.804: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 17:08:04.805: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.013259885s -Dec 14 17:08:04.805: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 17:08:06.805: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.01332138s -Dec 14 17:08:06.805: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 17:08:08.801: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.009399783s -Dec 14 17:08:08.801: INFO: The phase of Pod netserver-0 is Running (Ready = false) -Dec 14 17:08:10.802: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 12.009941843s -Dec 14 17:08:10.802: INFO: The phase of Pod netserver-0 is Running (Ready = true) -Dec 14 17:08:10.802: INFO: Pod "netserver-0" satisfied condition "running and ready" -Dec 14 17:08:10.808: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-3360" to be "running and ready" -Dec 14 17:08:10.813: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 5.019221ms -Dec 14 17:08:10.813: INFO: The phase of Pod netserver-1 is Running (Ready = true) -Dec 14 17:08:10.813: INFO: Pod "netserver-1" satisfied condition "running and ready" -Dec 14 17:08:10.817: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-3360" to be "running and ready" -Dec 14 17:08:10.821: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 4.128428ms -Dec 14 17:08:10.821: INFO: The phase of Pod netserver-2 is Running (Ready = false) -Dec 14 17:08:12.831: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 2.014328415s -Dec 14 17:08:12.831: INFO: The phase of Pod netserver-2 is Running (Ready = false) -Dec 14 17:08:14.831: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 4.013573981s -Dec 14 17:08:14.831: INFO: The phase of Pod netserver-2 is Running (Ready = false) -Dec 14 17:08:16.830: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 6.013132441s -Dec 14 17:08:16.830: INFO: The phase of Pod netserver-2 is Running (Ready = false) -Dec 14 17:08:18.827: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 8.010072426s -Dec 14 17:08:18.827: INFO: The phase of Pod netserver-2 is Running (Ready = false) -Dec 14 17:08:20.828: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 10.011039738s -Dec 14 17:08:20.828: INFO: The phase of Pod netserver-2 is Running (Ready = false) -Dec 14 17:08:22.828: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 12.011128847s -Dec 14 17:08:22.828: INFO: The phase of Pod netserver-2 is Running (Ready = false) -Dec 14 17:08:24.828: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 14.011325975s -Dec 14 17:08:24.828: INFO: The phase of Pod netserver-2 is Running (Ready = false) -Dec 14 17:08:26.827: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 16.010363698s -Dec 14 17:08:26.827: INFO: The phase of Pod netserver-2 is Running (Ready = true) -Dec 14 17:08:26.827: INFO: Pod "netserver-2" satisfied condition "running and ready" -STEP: Creating test pods 12/14/22 17:08:26.831 -Dec 14 17:08:26.861: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-3360" to be "running" -Dec 14 17:08:26.866: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 5.027596ms -Dec 14 17:08:28.874: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.012398791s -Dec 14 17:08:28.874: INFO: Pod "test-container-pod" satisfied condition "running" -Dec 14 17:08:28.879: INFO: Waiting up to 5m0s for pod "host-test-container-pod" in namespace "pod-network-test-3360" to be "running" -Dec 14 17:08:28.891: INFO: Pod "host-test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 12.161508ms -Dec 14 17:08:28.892: INFO: Pod "host-test-container-pod" satisfied condition "running" -Dec 14 17:08:28.896: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 -Dec 14 17:08:28.896: INFO: Going to poll 10.233.64.165 on port 8083 at least 0 times, with a maximum of 39 tries before failing -Dec 14 17:08:28.900: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.233.64.165:8083/hostName | grep -v '^\s*$'] Namespace:pod-network-test-3360 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 17:08:28.900: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 17:08:28.902: INFO: ExecWithOptions: Clientset creation -Dec 14 17:08:28.902: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-3360/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+--max-time+15+--connect-timeout+1+http%3A%2F%2F10.233.64.165%3A8083%2FhostName+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) -Dec 14 17:08:29.140: INFO: Found all 1 expected endpoints: [netserver-0] -Dec 14 17:08:29.140: INFO: Going to poll 10.233.65.163 on port 8083 at least 0 times, with a maximum of 39 tries before failing -Dec 14 17:08:29.151: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.233.65.163:8083/hostName | grep -v '^\s*$'] Namespace:pod-network-test-3360 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 17:08:29.151: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 17:08:29.153: INFO: ExecWithOptions: Clientset creation -Dec 14 17:08:29.153: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-3360/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+--max-time+15+--connect-timeout+1+http%3A%2F%2F10.233.65.163%3A8083%2FhostName+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) -Dec 14 17:08:29.306: INFO: Found all 1 expected endpoints: [netserver-1] -Dec 14 17:08:29.306: INFO: Going to poll 10.233.66.191 on port 8083 at least 0 times, with a maximum of 39 tries before failing -Dec 14 17:08:29.312: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.233.66.191:8083/hostName | grep -v '^\s*$'] Namespace:pod-network-test-3360 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 17:08:29.312: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 17:08:29.313: INFO: ExecWithOptions: Clientset creation -Dec 14 17:08:29.313: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-3360/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+--max-time+15+--connect-timeout+1+http%3A%2F%2F10.233.66.191%3A8083%2FhostName+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) -Dec 14 17:08:29.413: INFO: Found all 1 expected endpoints: [netserver-2] -[AfterEach] [sig-network] Networking +STEP: Creating a kubernetes client 02/25/23 11:40:10.037 +Feb 25 11:40:10.038: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubelet-test 02/25/23 11:40:10.041 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:40:10.077 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:40:10.08 +[BeforeEach] [sig-node] Kubelet + test/e2e/common/node/kubelet.go:41 +[BeforeEach] when scheduling a busybox command that always fails in a pod + test/e2e/common/node/kubelet.go:85 +[It] should have an terminated reason [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:110 +[AfterEach] [sig-node] Kubelet test/e2e/framework/framework.go:187 -Dec 14 17:08:29.414: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "pod-network-test-3360" for this suite. 12/14/22 17:08:29.423 -{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]","completed":341,"skipped":6366,"failed":0} +Feb 25 11:40:14.120: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubelet-test-8479" for this suite. 02/25/23 11:40:14.129 +{"msg":"PASSED [sig-node] Kubelet when scheduling a busybox command that always fails in a pod should have an terminated reason [NodeConformance] [Conformance]","completed":339,"skipped":6381,"failed":0} ------------------------------ -• [SLOW TEST] [30.716 seconds] -[sig-network] Networking -test/e2e/common/network/framework.go:23 - Granular Checks: Pods - test/e2e/common/network/networking.go:32 - should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:105 +• [4.104 seconds] +[sig-node] Kubelet +test/e2e/common/node/framework.go:23 + when scheduling a busybox command that always fails in a pod + test/e2e/common/node/kubelet.go:82 + should have an terminated reason [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:110 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Networking + [BeforeEach] [sig-node] Kubelet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:07:58.717 - Dec 14 17:07:58.717: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename pod-network-test 12/14/22 17:07:58.721 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:07:58.748 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:07:58.752 - [It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] - test/e2e/common/network/networking.go:105 - STEP: Performing setup for networking test in namespace pod-network-test-3360 12/14/22 17:07:58.755 - STEP: creating a selector 12/14/22 17:07:58.756 - STEP: Creating the service pods in kubernetes 12/14/22 17:07:58.756 - Dec 14 17:07:58.756: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable - Dec 14 17:07:58.792: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-3360" to be "running and ready" - Dec 14 17:07:58.795: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 3.469322ms - Dec 14 17:07:58.795: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 17:08:00.807: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 2.015012934s - Dec 14 17:08:00.807: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 17:08:02.803: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.011744195s - Dec 14 17:08:02.804: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 17:08:04.805: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.013259885s - Dec 14 17:08:04.805: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 17:08:06.805: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.01332138s - Dec 14 17:08:06.805: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 17:08:08.801: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.009399783s - Dec 14 17:08:08.801: INFO: The phase of Pod netserver-0 is Running (Ready = false) - Dec 14 17:08:10.802: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 12.009941843s - Dec 14 17:08:10.802: INFO: The phase of Pod netserver-0 is Running (Ready = true) - Dec 14 17:08:10.802: INFO: Pod "netserver-0" satisfied condition "running and ready" - Dec 14 17:08:10.808: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-3360" to be "running and ready" - Dec 14 17:08:10.813: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 5.019221ms - Dec 14 17:08:10.813: INFO: The phase of Pod netserver-1 is Running (Ready = true) - Dec 14 17:08:10.813: INFO: Pod "netserver-1" satisfied condition "running and ready" - Dec 14 17:08:10.817: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-3360" to be "running and ready" - Dec 14 17:08:10.821: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 4.128428ms - Dec 14 17:08:10.821: INFO: The phase of Pod netserver-2 is Running (Ready = false) - Dec 14 17:08:12.831: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 2.014328415s - Dec 14 17:08:12.831: INFO: The phase of Pod netserver-2 is Running (Ready = false) - Dec 14 17:08:14.831: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 4.013573981s - Dec 14 17:08:14.831: INFO: The phase of Pod netserver-2 is Running (Ready = false) - Dec 14 17:08:16.830: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 6.013132441s - Dec 14 17:08:16.830: INFO: The phase of Pod netserver-2 is Running (Ready = false) - Dec 14 17:08:18.827: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 8.010072426s - Dec 14 17:08:18.827: INFO: The phase of Pod netserver-2 is Running (Ready = false) - Dec 14 17:08:20.828: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 10.011039738s - Dec 14 17:08:20.828: INFO: The phase of Pod netserver-2 is Running (Ready = false) - Dec 14 17:08:22.828: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 12.011128847s - Dec 14 17:08:22.828: INFO: The phase of Pod netserver-2 is Running (Ready = false) - Dec 14 17:08:24.828: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=false. Elapsed: 14.011325975s - Dec 14 17:08:24.828: INFO: The phase of Pod netserver-2 is Running (Ready = false) - Dec 14 17:08:26.827: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 16.010363698s - Dec 14 17:08:26.827: INFO: The phase of Pod netserver-2 is Running (Ready = true) - Dec 14 17:08:26.827: INFO: Pod "netserver-2" satisfied condition "running and ready" - STEP: Creating test pods 12/14/22 17:08:26.831 - Dec 14 17:08:26.861: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-3360" to be "running" - Dec 14 17:08:26.866: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 5.027596ms - Dec 14 17:08:28.874: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 2.012398791s - Dec 14 17:08:28.874: INFO: Pod "test-container-pod" satisfied condition "running" - Dec 14 17:08:28.879: INFO: Waiting up to 5m0s for pod "host-test-container-pod" in namespace "pod-network-test-3360" to be "running" - Dec 14 17:08:28.891: INFO: Pod "host-test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 12.161508ms - Dec 14 17:08:28.892: INFO: Pod "host-test-container-pod" satisfied condition "running" - Dec 14 17:08:28.896: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 - Dec 14 17:08:28.896: INFO: Going to poll 10.233.64.165 on port 8083 at least 0 times, with a maximum of 39 tries before failing - Dec 14 17:08:28.900: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.233.64.165:8083/hostName | grep -v '^\s*$'] Namespace:pod-network-test-3360 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 17:08:28.900: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 17:08:28.902: INFO: ExecWithOptions: Clientset creation - Dec 14 17:08:28.902: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-3360/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+--max-time+15+--connect-timeout+1+http%3A%2F%2F10.233.64.165%3A8083%2FhostName+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) - Dec 14 17:08:29.140: INFO: Found all 1 expected endpoints: [netserver-0] - Dec 14 17:08:29.140: INFO: Going to poll 10.233.65.163 on port 8083 at least 0 times, with a maximum of 39 tries before failing - Dec 14 17:08:29.151: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.233.65.163:8083/hostName | grep -v '^\s*$'] Namespace:pod-network-test-3360 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 17:08:29.151: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 17:08:29.153: INFO: ExecWithOptions: Clientset creation - Dec 14 17:08:29.153: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-3360/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+--max-time+15+--connect-timeout+1+http%3A%2F%2F10.233.65.163%3A8083%2FhostName+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) - Dec 14 17:08:29.306: INFO: Found all 1 expected endpoints: [netserver-1] - Dec 14 17:08:29.306: INFO: Going to poll 10.233.66.191 on port 8083 at least 0 times, with a maximum of 39 tries before failing - Dec 14 17:08:29.312: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.233.66.191:8083/hostName | grep -v '^\s*$'] Namespace:pod-network-test-3360 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 17:08:29.312: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 17:08:29.313: INFO: ExecWithOptions: Clientset creation - Dec 14 17:08:29.313: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-3360/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+--max-time+15+--connect-timeout+1+http%3A%2F%2F10.233.66.191%3A8083%2FhostName+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) - Dec 14 17:08:29.413: INFO: Found all 1 expected endpoints: [netserver-2] - [AfterEach] [sig-network] Networking + STEP: Creating a kubernetes client 02/25/23 11:40:10.037 + Feb 25 11:40:10.038: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubelet-test 02/25/23 11:40:10.041 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:40:10.077 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:40:10.08 + [BeforeEach] [sig-node] Kubelet + test/e2e/common/node/kubelet.go:41 + [BeforeEach] when scheduling a busybox command that always fails in a pod + test/e2e/common/node/kubelet.go:85 + [It] should have an terminated reason [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:110 + [AfterEach] [sig-node] Kubelet test/e2e/framework/framework.go:187 - Dec 14 17:08:29.414: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "pod-network-test-3360" for this suite. 12/14/22 17:08:29.423 + Feb 25 11:40:14.120: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubelet-test-8479" for this suite. 02/25/23 11:40:14.129 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-network] EndpointSlice - should create Endpoints and EndpointSlices for Pods matching a Service [Conformance] - test/e2e/network/endpointslice.go:204 -[BeforeEach] [sig-network] EndpointSlice +[sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] + Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] + test/e2e/apps/statefulset.go:695 +[BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:08:29.437 -Dec 14 17:08:29.437: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename endpointslice 12/14/22 17:08:29.441 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:08:29.471 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:08:29.474 -[BeforeEach] [sig-network] EndpointSlice - test/e2e/network/endpointslice.go:51 -[It] should create Endpoints and EndpointSlices for Pods matching a Service [Conformance] - test/e2e/network/endpointslice.go:204 -STEP: referencing a single matching pod 12/14/22 17:08:34.623 -STEP: referencing matching pods with named port 12/14/22 17:08:39.64 -STEP: creating empty Endpoints and EndpointSlices for no matching Pods 12/14/22 17:08:44.658 -STEP: recreating EndpointSlices after they've been deleted 12/14/22 17:08:49.67 -Dec 14 17:08:49.713: INFO: EndpointSlice for Service endpointslice-7094/example-named-port not found -[AfterEach] [sig-network] EndpointSlice +STEP: Creating a kubernetes client 02/25/23 11:40:14.163 +Feb 25 11:40:14.164: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename statefulset 02/25/23 11:40:14.169 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:40:14.203 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:40:14.21 +[BeforeEach] [sig-apps] StatefulSet + test/e2e/apps/statefulset.go:96 +[BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:111 +STEP: Creating service test in namespace statefulset-3714 02/25/23 11:40:14.214 +[It] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] + test/e2e/apps/statefulset.go:695 +STEP: Creating stateful set ss in namespace statefulset-3714 02/25/23 11:40:14.223 +STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-3714 02/25/23 11:40:14.234 +Feb 25 11:40:14.241: INFO: Found 0 stateful pods, waiting for 1 +Feb 25 11:40:24.249: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true +STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod 02/25/23 11:40:24.249 +Feb 25 11:40:24.255: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-3714 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Feb 25 11:40:24.534: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Feb 25 11:40:24.534: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Feb 25 11:40:24.534: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +Feb 25 11:40:24.541: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true +Feb 25 11:40:34.547: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false +Feb 25 11:40:34.547: INFO: Waiting for statefulset status.replicas updated to 0 +Feb 25 11:40:34.572: INFO: POD NODE PHASE GRACE CONDITIONS +Feb 25 11:40:34.572: INFO: ss-0 diek3dua9gay-3 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC }] +Feb 25 11:40:34.572: INFO: +Feb 25 11:40:34.572: INFO: StatefulSet ss has not reached scale 3, at 1 +Feb 25 11:40:35.583: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.993723321s +Feb 25 11:40:36.590: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.982974218s +Feb 25 11:40:37.600: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.975962496s +Feb 25 11:40:38.608: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.965137772s +Feb 25 11:40:39.618: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.957952831s +Feb 25 11:40:40.627: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.946605827s +Feb 25 11:40:41.634: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.938520345s +Feb 25 11:40:42.645: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.931320374s +Feb 25 11:40:43.653: INFO: Verifying statefulset ss doesn't scale past 3 for another 921.192738ms +STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-3714 02/25/23 11:40:44.653 +Feb 25 11:40:44.660: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-3714 exec ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' +Feb 25 11:40:45.071: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" +Feb 25 11:40:45.072: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" +Feb 25 11:40:45.072: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + +Feb 25 11:40:45.073: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-3714 exec ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' +Feb 25 11:40:45.476: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n" +Feb 25 11:40:45.476: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" +Feb 25 11:40:45.476: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + +Feb 25 11:40:45.478: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-3714 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' +Feb 25 11:40:45.763: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n" +Feb 25 11:40:45.763: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" +Feb 25 11:40:45.763: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + +Feb 25 11:40:45.769: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true +Feb 25 11:40:45.769: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true +Feb 25 11:40:45.769: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true +STEP: Scale down will not halt with unhealthy stateful pod 02/25/23 11:40:45.769 +Feb 25 11:40:45.776: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-3714 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Feb 25 11:40:46.263: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Feb 25 11:40:46.263: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Feb 25 11:40:46.263: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +Feb 25 11:40:46.263: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-3714 exec ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Feb 25 11:40:46.550: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Feb 25 11:40:46.550: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Feb 25 11:40:46.550: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +Feb 25 11:40:46.550: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-3714 exec ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' +Feb 25 11:40:46.842: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" +Feb 25 11:40:46.842: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" +Feb 25 11:40:46.842: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + +Feb 25 11:40:46.842: INFO: Waiting for statefulset status.replicas updated to 0 +Feb 25 11:40:46.850: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 3 +Feb 25 11:40:56.876: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false +Feb 25 11:40:56.876: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false +Feb 25 11:40:56.876: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false +Feb 25 11:40:56.951: INFO: POD NODE PHASE GRACE CONDITIONS +Feb 25 11:40:56.951: INFO: ss-0 diek3dua9gay-3 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC }] +Feb 25 11:40:56.952: INFO: ss-1 diek3dua9gay-2 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC }] +Feb 25 11:40:56.952: INFO: ss-2 diek3dua9gay-1 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC }] +Feb 25 11:40:56.952: INFO: +Feb 25 11:40:56.952: INFO: StatefulSet ss has not reached scale 0, at 3 +Feb 25 11:40:57.959: INFO: POD NODE PHASE GRACE CONDITIONS +Feb 25 11:40:57.959: INFO: ss-0 diek3dua9gay-3 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC }] +Feb 25 11:40:57.960: INFO: ss-1 diek3dua9gay-2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC }] +Feb 25 11:40:57.960: INFO: ss-2 diek3dua9gay-1 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC }] +Feb 25 11:40:57.960: INFO: +Feb 25 11:40:57.960: INFO: StatefulSet ss has not reached scale 0, at 3 +Feb 25 11:40:58.967: INFO: POD NODE PHASE GRACE CONDITIONS +Feb 25 11:40:58.967: INFO: ss-0 diek3dua9gay-3 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC }] +Feb 25 11:40:58.967: INFO: +Feb 25 11:40:58.967: INFO: StatefulSet ss has not reached scale 0, at 1 +Feb 25 11:40:59.974: INFO: Verifying statefulset ss doesn't scale past 0 for another 6.975473895s +Feb 25 11:41:00.980: INFO: Verifying statefulset ss doesn't scale past 0 for another 5.969545855s +Feb 25 11:41:01.986: INFO: Verifying statefulset ss doesn't scale past 0 for another 4.96317748s +Feb 25 11:41:02.992: INFO: Verifying statefulset ss doesn't scale past 0 for another 3.957250126s +Feb 25 11:41:03.999: INFO: Verifying statefulset ss doesn't scale past 0 for another 2.951228422s +Feb 25 11:41:05.005: INFO: Verifying statefulset ss doesn't scale past 0 for another 1.944204732s +Feb 25 11:41:06.012: INFO: Verifying statefulset ss doesn't scale past 0 for another 938.240139ms +STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-3714 02/25/23 11:41:07.013 +Feb 25 11:41:07.020: INFO: Scaling statefulset ss to 0 +Feb 25 11:41:07.036: INFO: Waiting for statefulset status.replicas updated to 0 +[AfterEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:122 +Feb 25 11:41:07.045: INFO: Deleting all statefulset in ns statefulset-3714 +Feb 25 11:41:07.052: INFO: Scaling statefulset ss to 0 +Feb 25 11:41:07.071: INFO: Waiting for statefulset status.replicas updated to 0 +Feb 25 11:41:07.075: INFO: Deleting statefulset ss +[AfterEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:187 -Dec 14 17:08:59.732: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "endpointslice-7094" for this suite. 12/14/22 17:08:59.743 -{"msg":"PASSED [sig-network] EndpointSlice should create Endpoints and EndpointSlices for Pods matching a Service [Conformance]","completed":342,"skipped":6369,"failed":0} +Feb 25 11:41:07.093: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "statefulset-3714" for this suite. 02/25/23 11:41:07.1 +{"msg":"PASSED [sig-apps] StatefulSet Basic StatefulSet functionality [StatefulSetBasic] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance]","completed":340,"skipped":6427,"failed":0} ------------------------------ -• [SLOW TEST] [30.324 seconds] -[sig-network] EndpointSlice -test/e2e/network/common/framework.go:23 - should create Endpoints and EndpointSlices for Pods matching a Service [Conformance] - test/e2e/network/endpointslice.go:204 +• [SLOW TEST] [52.946 seconds] +[sig-apps] StatefulSet +test/e2e/apps/framework.go:23 + Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:101 + Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] + test/e2e/apps/statefulset.go:695 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] EndpointSlice + [BeforeEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:08:29.437 - Dec 14 17:08:29.437: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename endpointslice 12/14/22 17:08:29.441 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:08:29.471 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:08:29.474 - [BeforeEach] [sig-network] EndpointSlice - test/e2e/network/endpointslice.go:51 - [It] should create Endpoints and EndpointSlices for Pods matching a Service [Conformance] - test/e2e/network/endpointslice.go:204 - STEP: referencing a single matching pod 12/14/22 17:08:34.623 - STEP: referencing matching pods with named port 12/14/22 17:08:39.64 - STEP: creating empty Endpoints and EndpointSlices for no matching Pods 12/14/22 17:08:44.658 - STEP: recreating EndpointSlices after they've been deleted 12/14/22 17:08:49.67 - Dec 14 17:08:49.713: INFO: EndpointSlice for Service endpointslice-7094/example-named-port not found - [AfterEach] [sig-network] EndpointSlice + STEP: Creating a kubernetes client 02/25/23 11:40:14.163 + Feb 25 11:40:14.164: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename statefulset 02/25/23 11:40:14.169 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:40:14.203 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:40:14.21 + [BeforeEach] [sig-apps] StatefulSet + test/e2e/apps/statefulset.go:96 + [BeforeEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:111 + STEP: Creating service test in namespace statefulset-3714 02/25/23 11:40:14.214 + [It] Burst scaling should run to completion even with unhealthy pods [Slow] [Conformance] + test/e2e/apps/statefulset.go:695 + STEP: Creating stateful set ss in namespace statefulset-3714 02/25/23 11:40:14.223 + STEP: Waiting until all stateful set ss replicas will be running in namespace statefulset-3714 02/25/23 11:40:14.234 + Feb 25 11:40:14.241: INFO: Found 0 stateful pods, waiting for 1 + Feb 25 11:40:24.249: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true + STEP: Confirming that stateful set scale up will not halt with unhealthy stateful pod 02/25/23 11:40:24.249 + Feb 25 11:40:24.255: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-3714 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' + Feb 25 11:40:24.534: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" + Feb 25 11:40:24.534: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" + Feb 25 11:40:24.534: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + + Feb 25 11:40:24.541: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=true + Feb 25 11:40:34.547: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false + Feb 25 11:40:34.547: INFO: Waiting for statefulset status.replicas updated to 0 + Feb 25 11:40:34.572: INFO: POD NODE PHASE GRACE CONDITIONS + Feb 25 11:40:34.572: INFO: ss-0 diek3dua9gay-3 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:25 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC }] + Feb 25 11:40:34.572: INFO: + Feb 25 11:40:34.572: INFO: StatefulSet ss has not reached scale 3, at 1 + Feb 25 11:40:35.583: INFO: Verifying statefulset ss doesn't scale past 3 for another 8.993723321s + Feb 25 11:40:36.590: INFO: Verifying statefulset ss doesn't scale past 3 for another 7.982974218s + Feb 25 11:40:37.600: INFO: Verifying statefulset ss doesn't scale past 3 for another 6.975962496s + Feb 25 11:40:38.608: INFO: Verifying statefulset ss doesn't scale past 3 for another 5.965137772s + Feb 25 11:40:39.618: INFO: Verifying statefulset ss doesn't scale past 3 for another 4.957952831s + Feb 25 11:40:40.627: INFO: Verifying statefulset ss doesn't scale past 3 for another 3.946605827s + Feb 25 11:40:41.634: INFO: Verifying statefulset ss doesn't scale past 3 for another 2.938520345s + Feb 25 11:40:42.645: INFO: Verifying statefulset ss doesn't scale past 3 for another 1.931320374s + Feb 25 11:40:43.653: INFO: Verifying statefulset ss doesn't scale past 3 for another 921.192738ms + STEP: Scaling up stateful set ss to 3 replicas and waiting until all of them will be running in namespace statefulset-3714 02/25/23 11:40:44.653 + Feb 25 11:40:44.660: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-3714 exec ss-0 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' + Feb 25 11:40:45.071: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\n" + Feb 25 11:40:45.072: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" + Feb 25 11:40:45.072: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-0: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + + Feb 25 11:40:45.073: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-3714 exec ss-1 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' + Feb 25 11:40:45.476: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n" + Feb 25 11:40:45.476: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" + Feb 25 11:40:45.476: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-1: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + + Feb 25 11:40:45.478: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-3714 exec ss-2 -- /bin/sh -x -c mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true' + Feb 25 11:40:45.763: INFO: stderr: "+ mv -v /tmp/index.html /usr/local/apache2/htdocs/\nmv: can't rename '/tmp/index.html': No such file or directory\n+ true\n" + Feb 25 11:40:45.763: INFO: stdout: "'/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html'\n" + Feb 25 11:40:45.763: INFO: stdout of mv -v /tmp/index.html /usr/local/apache2/htdocs/ || true on ss-2: '/tmp/index.html' -> '/usr/local/apache2/htdocs/index.html' + + Feb 25 11:40:45.769: INFO: Waiting for pod ss-0 to enter Running - Ready=true, currently Running - Ready=true + Feb 25 11:40:45.769: INFO: Waiting for pod ss-1 to enter Running - Ready=true, currently Running - Ready=true + Feb 25 11:40:45.769: INFO: Waiting for pod ss-2 to enter Running - Ready=true, currently Running - Ready=true + STEP: Scale down will not halt with unhealthy stateful pod 02/25/23 11:40:45.769 + Feb 25 11:40:45.776: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-3714 exec ss-0 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' + Feb 25 11:40:46.263: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" + Feb 25 11:40:46.263: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" + Feb 25 11:40:46.263: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-0: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + + Feb 25 11:40:46.263: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-3714 exec ss-1 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' + Feb 25 11:40:46.550: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" + Feb 25 11:40:46.550: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" + Feb 25 11:40:46.550: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-1: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + + Feb 25 11:40:46.550: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=statefulset-3714 exec ss-2 -- /bin/sh -x -c mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true' + Feb 25 11:40:46.842: INFO: stderr: "+ mv -v /usr/local/apache2/htdocs/index.html /tmp/\n" + Feb 25 11:40:46.842: INFO: stdout: "'/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html'\n" + Feb 25 11:40:46.842: INFO: stdout of mv -v /usr/local/apache2/htdocs/index.html /tmp/ || true on ss-2: '/usr/local/apache2/htdocs/index.html' -> '/tmp/index.html' + + Feb 25 11:40:46.842: INFO: Waiting for statefulset status.replicas updated to 0 + Feb 25 11:40:46.850: INFO: Waiting for stateful set status.readyReplicas to become 0, currently 3 + Feb 25 11:40:56.876: INFO: Waiting for pod ss-0 to enter Running - Ready=false, currently Running - Ready=false + Feb 25 11:40:56.876: INFO: Waiting for pod ss-1 to enter Running - Ready=false, currently Running - Ready=false + Feb 25 11:40:56.876: INFO: Waiting for pod ss-2 to enter Running - Ready=false, currently Running - Ready=false + Feb 25 11:40:56.951: INFO: POD NODE PHASE GRACE CONDITIONS + Feb 25 11:40:56.951: INFO: ss-0 diek3dua9gay-3 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC }] + Feb 25 11:40:56.952: INFO: ss-1 diek3dua9gay-2 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC }] + Feb 25 11:40:56.952: INFO: ss-2 diek3dua9gay-1 Running [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC }] + Feb 25 11:40:56.952: INFO: + Feb 25 11:40:56.952: INFO: StatefulSet ss has not reached scale 0, at 3 + Feb 25 11:40:57.959: INFO: POD NODE PHASE GRACE CONDITIONS + Feb 25 11:40:57.959: INFO: ss-0 diek3dua9gay-3 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC }] + Feb 25 11:40:57.960: INFO: ss-1 diek3dua9gay-2 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC }] + Feb 25 11:40:57.960: INFO: ss-2 diek3dua9gay-1 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:34 +0000 UTC }] + Feb 25 11:40:57.960: INFO: + Feb 25 11:40:57.960: INFO: StatefulSet ss has not reached scale 0, at 3 + Feb 25 11:40:58.967: INFO: POD NODE PHASE GRACE CONDITIONS + Feb 25 11:40:58.967: INFO: ss-0 diek3dua9gay-3 Running 30s [{Initialized True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC } {Ready False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {ContainersReady False 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:47 +0000 UTC ContainersNotReady containers with unready status: [webserver]} {PodScheduled True 0001-01-01 00:00:00 +0000 UTC 2023-02-25 11:40:14 +0000 UTC }] + Feb 25 11:40:58.967: INFO: + Feb 25 11:40:58.967: INFO: StatefulSet ss has not reached scale 0, at 1 + Feb 25 11:40:59.974: INFO: Verifying statefulset ss doesn't scale past 0 for another 6.975473895s + Feb 25 11:41:00.980: INFO: Verifying statefulset ss doesn't scale past 0 for another 5.969545855s + Feb 25 11:41:01.986: INFO: Verifying statefulset ss doesn't scale past 0 for another 4.96317748s + Feb 25 11:41:02.992: INFO: Verifying statefulset ss doesn't scale past 0 for another 3.957250126s + Feb 25 11:41:03.999: INFO: Verifying statefulset ss doesn't scale past 0 for another 2.951228422s + Feb 25 11:41:05.005: INFO: Verifying statefulset ss doesn't scale past 0 for another 1.944204732s + Feb 25 11:41:06.012: INFO: Verifying statefulset ss doesn't scale past 0 for another 938.240139ms + STEP: Scaling down stateful set ss to 0 replicas and waiting until none of pods will run in namespacestatefulset-3714 02/25/23 11:41:07.013 + Feb 25 11:41:07.020: INFO: Scaling statefulset ss to 0 + Feb 25 11:41:07.036: INFO: Waiting for statefulset status.replicas updated to 0 + [AfterEach] Basic StatefulSet functionality [StatefulSetBasic] + test/e2e/apps/statefulset.go:122 + Feb 25 11:41:07.045: INFO: Deleting all statefulset in ns statefulset-3714 + Feb 25 11:41:07.052: INFO: Scaling statefulset ss to 0 + Feb 25 11:41:07.071: INFO: Waiting for statefulset status.replicas updated to 0 + Feb 25 11:41:07.075: INFO: Deleting statefulset ss + [AfterEach] [sig-apps] StatefulSet test/e2e/framework/framework.go:187 - Dec 14 17:08:59.732: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "endpointslice-7094" for this suite. 12/14/22 17:08:59.743 + Feb 25 11:41:07.093: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "statefulset-3714" for this suite. 02/25/23 11:41:07.1 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SS ------------------------------ -[sig-storage] Secrets - should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:78 -[BeforeEach] [sig-storage] Secrets +[sig-network] EndpointSliceMirroring + should mirror a custom Endpoints resource through create update and delete [Conformance] + test/e2e/network/endpointslicemirroring.go:53 +[BeforeEach] [sig-network] EndpointSliceMirroring test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:08:59.768 -Dec 14 17:08:59.769: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename secrets 12/14/22 17:08:59.772 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:08:59.816 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:08:59.82 -[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:78 -STEP: Creating secret with name secret-test-map-2188cc5a-28f4-41f8-8d06-fd99589f604b 12/14/22 17:08:59.823 -STEP: Creating a pod to test consume secrets 12/14/22 17:08:59.831 -Dec 14 17:08:59.866: INFO: Waiting up to 5m0s for pod "pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45" in namespace "secrets-3188" to be "Succeeded or Failed" -Dec 14 17:08:59.871: INFO: Pod "pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45": Phase="Pending", Reason="", readiness=false. Elapsed: 4.528403ms -Dec 14 17:09:01.875: INFO: Pod "pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008943818s -Dec 14 17:09:03.878: INFO: Pod "pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011778446s -Dec 14 17:09:05.887: INFO: Pod "pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.02100275s -STEP: Saw pod success 12/14/22 17:09:05.887 -Dec 14 17:09:05.890: INFO: Pod "pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45" satisfied condition "Succeeded or Failed" -Dec 14 17:09:05.897: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45 container secret-volume-test: -STEP: delete the pod 12/14/22 17:09:06.304 -Dec 14 17:09:06.336: INFO: Waiting for pod pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45 to disappear -Dec 14 17:09:06.344: INFO: Pod pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45 no longer exists -[AfterEach] [sig-storage] Secrets +STEP: Creating a kubernetes client 02/25/23 11:41:07.111 +Feb 25 11:41:07.112: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename endpointslicemirroring 02/25/23 11:41:07.114 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:07.135 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:07.14 +[BeforeEach] [sig-network] EndpointSliceMirroring + test/e2e/network/endpointslicemirroring.go:41 +[It] should mirror a custom Endpoints resource through create update and delete [Conformance] + test/e2e/network/endpointslicemirroring.go:53 +STEP: mirroring a new custom Endpoint 02/25/23 11:41:07.158 +Feb 25 11:41:07.172: INFO: Waiting for at least 1 EndpointSlice to exist, got 0 +STEP: mirroring an update to a custom Endpoint 02/25/23 11:41:09.212 +Feb 25 11:41:09.251: INFO: Expected EndpointSlice to have 10.2.3.4 as address, got 10.1.2.3 +STEP: mirroring deletion of a custom Endpoint 02/25/23 11:41:11.258 +Feb 25 11:41:11.274: INFO: Waiting for 0 EndpointSlices to exist, got 1 +[AfterEach] [sig-network] EndpointSliceMirroring test/e2e/framework/framework.go:187 -Dec 14 17:09:06.344: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-3188" for this suite. 12/14/22 17:09:06.351 -{"msg":"PASSED [sig-storage] Secrets should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","completed":343,"skipped":6408,"failed":0} +Feb 25 11:41:13.283: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "endpointslicemirroring-8343" for this suite. 02/25/23 11:41:13.297 +{"msg":"PASSED [sig-network] EndpointSliceMirroring should mirror a custom Endpoints resource through create update and delete [Conformance]","completed":341,"skipped":6429,"failed":0} ------------------------------ -• [SLOW TEST] [6.595 seconds] -[sig-storage] Secrets -test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:78 +• [SLOW TEST] [6.197 seconds] +[sig-network] EndpointSliceMirroring +test/e2e/network/common/framework.go:23 + should mirror a custom Endpoints resource through create update and delete [Conformance] + test/e2e/network/endpointslicemirroring.go:53 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Secrets + [BeforeEach] [sig-network] EndpointSliceMirroring test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:08:59.768 - Dec 14 17:08:59.769: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename secrets 12/14/22 17:08:59.772 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:08:59.816 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:08:59.82 - [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] - test/e2e/common/storage/secrets_volume.go:78 - STEP: Creating secret with name secret-test-map-2188cc5a-28f4-41f8-8d06-fd99589f604b 12/14/22 17:08:59.823 - STEP: Creating a pod to test consume secrets 12/14/22 17:08:59.831 - Dec 14 17:08:59.866: INFO: Waiting up to 5m0s for pod "pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45" in namespace "secrets-3188" to be "Succeeded or Failed" - Dec 14 17:08:59.871: INFO: Pod "pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45": Phase="Pending", Reason="", readiness=false. Elapsed: 4.528403ms - Dec 14 17:09:01.875: INFO: Pod "pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45": Phase="Pending", Reason="", readiness=false. Elapsed: 2.008943818s - Dec 14 17:09:03.878: INFO: Pod "pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011778446s - Dec 14 17:09:05.887: INFO: Pod "pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.02100275s - STEP: Saw pod success 12/14/22 17:09:05.887 - Dec 14 17:09:05.890: INFO: Pod "pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45" satisfied condition "Succeeded or Failed" - Dec 14 17:09:05.897: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45 container secret-volume-test: - STEP: delete the pod 12/14/22 17:09:06.304 - Dec 14 17:09:06.336: INFO: Waiting for pod pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45 to disappear - Dec 14 17:09:06.344: INFO: Pod pod-secrets-fe7b6b1e-d0f4-4738-ad68-0c2c2bd6be45 no longer exists - [AfterEach] [sig-storage] Secrets + STEP: Creating a kubernetes client 02/25/23 11:41:07.111 + Feb 25 11:41:07.112: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename endpointslicemirroring 02/25/23 11:41:07.114 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:07.135 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:07.14 + [BeforeEach] [sig-network] EndpointSliceMirroring + test/e2e/network/endpointslicemirroring.go:41 + [It] should mirror a custom Endpoints resource through create update and delete [Conformance] + test/e2e/network/endpointslicemirroring.go:53 + STEP: mirroring a new custom Endpoint 02/25/23 11:41:07.158 + Feb 25 11:41:07.172: INFO: Waiting for at least 1 EndpointSlice to exist, got 0 + STEP: mirroring an update to a custom Endpoint 02/25/23 11:41:09.212 + Feb 25 11:41:09.251: INFO: Expected EndpointSlice to have 10.2.3.4 as address, got 10.1.2.3 + STEP: mirroring deletion of a custom Endpoint 02/25/23 11:41:11.258 + Feb 25 11:41:11.274: INFO: Waiting for 0 EndpointSlices to exist, got 1 + [AfterEach] [sig-network] EndpointSliceMirroring test/e2e/framework/framework.go:187 - Dec 14 17:09:06.344: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "secrets-3188" for this suite. 12/14/22 17:09:06.351 + Feb 25 11:41:13.283: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "endpointslicemirroring-8343" for this suite. 02/25/23 11:41:13.297 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------- -[sig-node] Security Context When creating a pod with readOnlyRootFilesystem - should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:485 -[BeforeEach] [sig-node] Security Context +[sig-instrumentation] Events + should manage the lifecycle of an event [Conformance] + test/e2e/instrumentation/core_events.go:57 +[BeforeEach] [sig-instrumentation] Events test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:09:06.369 -Dec 14 17:09:06.369: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename security-context-test 12/14/22 17:09:06.37 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:06.4 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:06.405 -[BeforeEach] [sig-node] Security Context - test/e2e/common/node/security_context.go:49 -[It] should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:485 -Dec 14 17:09:06.433: INFO: Waiting up to 5m0s for pod "busybox-readonly-false-9c382cb6-acba-46a9-9267-ff3088559f1f" in namespace "security-context-test-7140" to be "Succeeded or Failed" -Dec 14 17:09:06.438: INFO: Pod "busybox-readonly-false-9c382cb6-acba-46a9-9267-ff3088559f1f": Phase="Pending", Reason="", readiness=false. Elapsed: 5.638082ms -Dec 14 17:09:08.445: INFO: Pod "busybox-readonly-false-9c382cb6-acba-46a9-9267-ff3088559f1f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012466364s -Dec 14 17:09:10.445: INFO: Pod "busybox-readonly-false-9c382cb6-acba-46a9-9267-ff3088559f1f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012226747s -Dec 14 17:09:10.445: INFO: Pod "busybox-readonly-false-9c382cb6-acba-46a9-9267-ff3088559f1f" satisfied condition "Succeeded or Failed" -[AfterEach] [sig-node] Security Context +STEP: Creating a kubernetes client 02/25/23 11:41:13.311 +Feb 25 11:41:13.311: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename events 02/25/23 11:41:13.314 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:13.343 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:13.349 +[It] should manage the lifecycle of an event [Conformance] + test/e2e/instrumentation/core_events.go:57 +STEP: creating a test event 02/25/23 11:41:13.354 +STEP: listing all events in all namespaces 02/25/23 11:41:13.362 +STEP: patching the test event 02/25/23 11:41:13.371 +STEP: fetching the test event 02/25/23 11:41:13.383 +STEP: updating the test event 02/25/23 11:41:13.388 +STEP: getting the test event 02/25/23 11:41:13.41 +STEP: deleting the test event 02/25/23 11:41:13.415 +STEP: listing all events in all namespaces 02/25/23 11:41:13.432 +[AfterEach] [sig-instrumentation] Events test/e2e/framework/framework.go:187 -Dec 14 17:09:10.445: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "security-context-test-7140" for this suite. 12/14/22 17:09:10.452 -{"msg":"PASSED [sig-node] Security Context When creating a pod with readOnlyRootFilesystem should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance]","completed":344,"skipped":6434,"failed":0} +Feb 25 11:41:13.441: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "events-8542" for this suite. 02/25/23 11:41:13.455 +{"msg":"PASSED [sig-instrumentation] Events should manage the lifecycle of an event [Conformance]","completed":342,"skipped":6429,"failed":0} ------------------------------ -• [4.091 seconds] -[sig-node] Security Context -test/e2e/common/node/framework.go:23 - When creating a pod with readOnlyRootFilesystem - test/e2e/common/node/security_context.go:429 - should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:485 +• [0.156 seconds] +[sig-instrumentation] Events +test/e2e/instrumentation/common/framework.go:23 + should manage the lifecycle of an event [Conformance] + test/e2e/instrumentation/core_events.go:57 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Security Context + [BeforeEach] [sig-instrumentation] Events test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:09:06.369 - Dec 14 17:09:06.369: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename security-context-test 12/14/22 17:09:06.37 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:06.4 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:06.405 - [BeforeEach] [sig-node] Security Context - test/e2e/common/node/security_context.go:49 - [It] should run the container with writable rootfs when readOnlyRootFilesystem=false [NodeConformance] [Conformance] - test/e2e/common/node/security_context.go:485 - Dec 14 17:09:06.433: INFO: Waiting up to 5m0s for pod "busybox-readonly-false-9c382cb6-acba-46a9-9267-ff3088559f1f" in namespace "security-context-test-7140" to be "Succeeded or Failed" - Dec 14 17:09:06.438: INFO: Pod "busybox-readonly-false-9c382cb6-acba-46a9-9267-ff3088559f1f": Phase="Pending", Reason="", readiness=false. Elapsed: 5.638082ms - Dec 14 17:09:08.445: INFO: Pod "busybox-readonly-false-9c382cb6-acba-46a9-9267-ff3088559f1f": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012466364s - Dec 14 17:09:10.445: INFO: Pod "busybox-readonly-false-9c382cb6-acba-46a9-9267-ff3088559f1f": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.012226747s - Dec 14 17:09:10.445: INFO: Pod "busybox-readonly-false-9c382cb6-acba-46a9-9267-ff3088559f1f" satisfied condition "Succeeded or Failed" - [AfterEach] [sig-node] Security Context + STEP: Creating a kubernetes client 02/25/23 11:41:13.311 + Feb 25 11:41:13.311: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename events 02/25/23 11:41:13.314 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:13.343 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:13.349 + [It] should manage the lifecycle of an event [Conformance] + test/e2e/instrumentation/core_events.go:57 + STEP: creating a test event 02/25/23 11:41:13.354 + STEP: listing all events in all namespaces 02/25/23 11:41:13.362 + STEP: patching the test event 02/25/23 11:41:13.371 + STEP: fetching the test event 02/25/23 11:41:13.383 + STEP: updating the test event 02/25/23 11:41:13.388 + STEP: getting the test event 02/25/23 11:41:13.41 + STEP: deleting the test event 02/25/23 11:41:13.415 + STEP: listing all events in all namespaces 02/25/23 11:41:13.432 + [AfterEach] [sig-instrumentation] Events test/e2e/framework/framework.go:187 - Dec 14 17:09:10.445: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "security-context-test-7140" for this suite. 12/14/22 17:09:10.452 + Feb 25 11:41:13.441: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "events-8542" for this suite. 02/25/23 11:41:13.455 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSS +SSSSSSSSSSS ------------------------------ -[sig-apps] Job - should apply changes to a job status [Conformance] - test/e2e/apps/job.go:464 -[BeforeEach] [sig-apps] Job +[sig-apps] ReplicaSet + Replicaset should have a working scale subresource [Conformance] + test/e2e/apps/replica_set.go:143 +[BeforeEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:09:10.462 -Dec 14 17:09:10.463: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename job 12/14/22 17:09:10.465 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:10.486 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:10.489 -[It] should apply changes to a job status [Conformance] - test/e2e/apps/job.go:464 -STEP: Creating a job 12/14/22 17:09:10.493 -STEP: Ensure pods equal to paralellism count is attached to the job 12/14/22 17:09:10.501 -STEP: patching /status 12/14/22 17:09:12.508 -STEP: updating /status 12/14/22 17:09:12.522 -STEP: get /status 12/14/22 17:09:12.564 -[AfterEach] [sig-apps] Job +STEP: Creating a kubernetes client 02/25/23 11:41:13.471 +Feb 25 11:41:13.472: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename replicaset 02/25/23 11:41:13.473 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:13.506 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:13.51 +[It] Replicaset should have a working scale subresource [Conformance] + test/e2e/apps/replica_set.go:143 +STEP: Creating replica set "test-rs" that asks for more than the allowed pod quota 02/25/23 11:41:13.513 +Feb 25 11:41:13.538: INFO: Pod name sample-pod: Found 0 pods out of 1 +Feb 25 11:41:18.546: INFO: Pod name sample-pod: Found 1 pods out of 1 +STEP: ensuring each pod is running 02/25/23 11:41:18.546 +STEP: getting scale subresource 02/25/23 11:41:18.546 +STEP: updating a scale subresource 02/25/23 11:41:18.551 +STEP: verifying the replicaset Spec.Replicas was modified 02/25/23 11:41:18.568 +STEP: Patch a scale subresource 02/25/23 11:41:18.574 +[AfterEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:187 -Dec 14 17:09:12.569: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "job-185" for this suite. 12/14/22 17:09:12.574 -{"msg":"PASSED [sig-apps] Job should apply changes to a job status [Conformance]","completed":345,"skipped":6449,"failed":0} +Feb 25 11:41:18.595: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "replicaset-5096" for this suite. 02/25/23 11:41:18.611 +{"msg":"PASSED [sig-apps] ReplicaSet Replicaset should have a working scale subresource [Conformance]","completed":343,"skipped":6440,"failed":0} ------------------------------ -• [2.119 seconds] -[sig-apps] Job +• [SLOW TEST] [5.155 seconds] +[sig-apps] ReplicaSet test/e2e/apps/framework.go:23 - should apply changes to a job status [Conformance] - test/e2e/apps/job.go:464 + Replicaset should have a working scale subresource [Conformance] + test/e2e/apps/replica_set.go:143 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Job + [BeforeEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:09:10.462 - Dec 14 17:09:10.463: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename job 12/14/22 17:09:10.465 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:10.486 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:10.489 - [It] should apply changes to a job status [Conformance] - test/e2e/apps/job.go:464 - STEP: Creating a job 12/14/22 17:09:10.493 - STEP: Ensure pods equal to paralellism count is attached to the job 12/14/22 17:09:10.501 - STEP: patching /status 12/14/22 17:09:12.508 - STEP: updating /status 12/14/22 17:09:12.522 - STEP: get /status 12/14/22 17:09:12.564 - [AfterEach] [sig-apps] Job + STEP: Creating a kubernetes client 02/25/23 11:41:13.471 + Feb 25 11:41:13.472: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename replicaset 02/25/23 11:41:13.473 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:13.506 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:13.51 + [It] Replicaset should have a working scale subresource [Conformance] + test/e2e/apps/replica_set.go:143 + STEP: Creating replica set "test-rs" that asks for more than the allowed pod quota 02/25/23 11:41:13.513 + Feb 25 11:41:13.538: INFO: Pod name sample-pod: Found 0 pods out of 1 + Feb 25 11:41:18.546: INFO: Pod name sample-pod: Found 1 pods out of 1 + STEP: ensuring each pod is running 02/25/23 11:41:18.546 + STEP: getting scale subresource 02/25/23 11:41:18.546 + STEP: updating a scale subresource 02/25/23 11:41:18.551 + STEP: verifying the replicaset Spec.Replicas was modified 02/25/23 11:41:18.568 + STEP: Patch a scale subresource 02/25/23 11:41:18.574 + [AfterEach] [sig-apps] ReplicaSet test/e2e/framework/framework.go:187 - Dec 14 17:09:12.569: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "job-185" for this suite. 12/14/22 17:09:12.574 + Feb 25 11:41:18.595: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "replicaset-5096" for this suite. 02/25/23 11:41:18.611 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSS +SSSSSSSSSSS ------------------------------ -[sig-storage] ConfigMap - updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:123 -[BeforeEach] [sig-storage] ConfigMap +[sig-scheduling] SchedulerPredicates [Serial] + validates that NodeSelector is respected if not matching [Conformance] + test/e2e/scheduling/predicates.go:438 +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:09:12.585 -Dec 14 17:09:12.585: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename configmap 12/14/22 17:09:12.588 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:12.614 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:12.618 -[It] updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:123 -STEP: Creating configMap with name configmap-test-upd-1fdd77b9-36e8-475f-8116-2f796284a1fa 12/14/22 17:09:12.628 -STEP: Creating the pod 12/14/22 17:09:12.634 -Dec 14 17:09:12.645: INFO: Waiting up to 5m0s for pod "pod-configmaps-4c016e0a-9f52-4195-9639-37c11eb2e069" in namespace "configmap-9236" to be "running and ready" -Dec 14 17:09:12.651: INFO: Pod "pod-configmaps-4c016e0a-9f52-4195-9639-37c11eb2e069": Phase="Pending", Reason="", readiness=false. Elapsed: 6.180836ms -Dec 14 17:09:12.651: INFO: The phase of Pod pod-configmaps-4c016e0a-9f52-4195-9639-37c11eb2e069 is Pending, waiting for it to be Running (with Ready = true) -Dec 14 17:09:14.667: INFO: Pod "pod-configmaps-4c016e0a-9f52-4195-9639-37c11eb2e069": Phase="Running", Reason="", readiness=true. Elapsed: 2.021789128s -Dec 14 17:09:14.667: INFO: The phase of Pod pod-configmaps-4c016e0a-9f52-4195-9639-37c11eb2e069 is Running (Ready = true) -Dec 14 17:09:14.667: INFO: Pod "pod-configmaps-4c016e0a-9f52-4195-9639-37c11eb2e069" satisfied condition "running and ready" -STEP: Updating configmap configmap-test-upd-1fdd77b9-36e8-475f-8116-2f796284a1fa 12/14/22 17:09:14.696 -STEP: waiting to observe update in volume 12/14/22 17:09:14.71 -[AfterEach] [sig-storage] ConfigMap +STEP: Creating a kubernetes client 02/25/23 11:41:18.652 +Feb 25 11:41:18.653: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename sched-pred 02/25/23 11:41:18.66 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:18.722 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:18.727 +[BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:92 +Feb 25 11:41:18.731: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready +Feb 25 11:41:18.749: INFO: Waiting for terminating namespaces to be deleted... +Feb 25 11:41:18.757: INFO: +Logging pods the apiserver thinks is on node diek3dua9gay-1 before test +Feb 25 11:41:18.776: INFO: echo-other-node-cd69fcf6b-hsnk9 from cilium-test started at 2023-02-25 09:38:06 +0000 UTC (2 container statuses recorded) +Feb 25 11:41:18.776: INFO: Container dns-test-server ready: true, restart count 0 +Feb 25 11:41:18.776: INFO: Container echo-other-node ready: true, restart count 0 +Feb 25 11:41:18.777: INFO: cilium-node-init-s84z5 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.777: INFO: Container node-init ready: true, restart count 0 +Feb 25 11:41:18.777: INFO: cilium-r4lc4 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.777: INFO: Container cilium-agent ready: true, restart count 0 +Feb 25 11:41:18.777: INFO: coredns-565d847f94-qqvf6 from kube-system started at 2023-02-25 08:49:06 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.777: INFO: Container coredns ready: true, restart count 0 +Feb 25 11:41:18.777: INFO: kube-addon-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.778: INFO: Container kube-addon-manager ready: true, restart count 0 +Feb 25 11:41:18.778: INFO: kube-apiserver-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:09 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.778: INFO: Container kube-apiserver ready: true, restart count 0 +Feb 25 11:41:18.778: INFO: kube-controller-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:49:55 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.778: INFO: Container kube-controller-manager ready: true, restart count 0 +Feb 25 11:41:18.778: INFO: kube-proxy-cm4qg from kube-system started at 2023-02-25 08:46:38 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.778: INFO: Container kube-proxy ready: true, restart count 0 +Feb 25 11:41:18.778: INFO: kube-scheduler-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:27 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.778: INFO: Container kube-scheduler ready: true, restart count 0 +Feb 25 11:41:18.778: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-wznrx from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 11:41:18.778: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 11:41:18.778: INFO: Container systemd-logs ready: true, restart count 0 +Feb 25 11:41:18.778: INFO: +Logging pods the apiserver thinks is on node diek3dua9gay-2 before test +Feb 25 11:41:18.803: INFO: cilium-9tkfg from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.803: INFO: Container cilium-agent ready: true, restart count 0 +Feb 25 11:41:18.803: INFO: cilium-node-init-8b7n2 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.803: INFO: Container node-init ready: true, restart count 0 +Feb 25 11:41:18.803: INFO: cilium-operator-59d7d77cdc-lppbd from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.803: INFO: Container cilium-operator ready: true, restart count 0 +Feb 25 11:41:18.803: INFO: coredns-565d847f94-5nh2n from kube-system started at 2023-02-25 08:48:51 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.803: INFO: Container coredns ready: true, restart count 0 +Feb 25 11:41:18.803: INFO: kube-addon-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.803: INFO: Container kube-addon-manager ready: true, restart count 0 +Feb 25 11:41:18.803: INFO: kube-apiserver-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.803: INFO: Container kube-apiserver ready: true, restart count 0 +Feb 25 11:41:18.803: INFO: kube-controller-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.803: INFO: Container kube-controller-manager ready: true, restart count 0 +Feb 25 11:41:18.803: INFO: kube-proxy-d66m7 from kube-system started at 2023-02-25 08:46:57 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.803: INFO: Container kube-proxy ready: true, restart count 0 +Feb 25 11:41:18.803: INFO: kube-scheduler-diek3dua9gay-2 from kube-system started at 2023-02-25 08:50:15 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.803: INFO: Container kube-scheduler ready: true, restart count 0 +Feb 25 11:41:18.803: INFO: test-rs-946fw from replicaset-5096 started at 2023-02-25 11:41:18 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.803: INFO: Container httpd ready: false, restart count 0 +Feb 25 11:41:18.803: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-mml8w from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 11:41:18.803: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 11:41:18.803: INFO: Container systemd-logs ready: true, restart count 0 +Feb 25 11:41:18.803: INFO: +Logging pods the apiserver thinks is on node diek3dua9gay-3 before test +Feb 25 11:41:18.827: INFO: client-7b78db77d5-27xbt from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.827: INFO: Container client ready: true, restart count 0 +Feb 25 11:41:18.827: INFO: client2-78f748dd67-skwz2 from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.827: INFO: Container client2 ready: true, restart count 0 +Feb 25 11:41:18.827: INFO: echo-same-node-85bc9b6b56-wqmfq from cilium-test started at 2023-02-25 09:57:46 +0000 UTC (2 container statuses recorded) +Feb 25 11:41:18.827: INFO: Container dns-test-server ready: true, restart count 0 +Feb 25 11:41:18.827: INFO: Container echo-same-node ready: true, restart count 0 +Feb 25 11:41:18.827: INFO: cilium-2cncz from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.828: INFO: Container cilium-agent ready: true, restart count 0 +Feb 25 11:41:18.828: INFO: cilium-node-init-979mb from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.828: INFO: Container node-init ready: true, restart count 0 +Feb 25 11:41:18.828: INFO: kube-proxy-22s9q from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.828: INFO: Container kube-proxy ready: true, restart count 0 +Feb 25 11:41:18.828: INFO: test-rs-2g7jr from replicaset-5096 started at 2023-02-25 11:41:13 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.828: INFO: Container httpd ready: true, restart count 0 +Feb 25 11:41:18.828: INFO: sonobuoy from sonobuoy started at 2023-02-25 09:45:14 +0000 UTC (1 container statuses recorded) +Feb 25 11:41:18.828: INFO: Container kube-sonobuoy ready: true, restart count 0 +Feb 25 11:41:18.828: INFO: sonobuoy-e2e-job-0b3c30975b5c4880 from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 11:41:18.828: INFO: Container e2e ready: true, restart count 0 +Feb 25 11:41:18.828: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 11:41:18.828: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-9pdjb from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) +Feb 25 11:41:18.828: INFO: Container sonobuoy-worker ready: true, restart count 0 +Feb 25 11:41:18.828: INFO: Container systemd-logs ready: true, restart count 0 +[It] validates that NodeSelector is respected if not matching [Conformance] + test/e2e/scheduling/predicates.go:438 +STEP: Trying to schedule Pod with nonempty NodeSelector. 02/25/23 11:41:18.829 +STEP: Considering event: +Type = [Warning], Name = [restricted-pod.17470e75b97347ce], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match Pod's node affinity/selector. preemption: 0/3 nodes are available: 3 Preemption is not helpful for scheduling.] 02/25/23 11:41:25.024 +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:187 -Dec 14 17:09:16.734: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-9236" for this suite. 12/14/22 17:09:16.74 -{"msg":"PASSED [sig-storage] ConfigMap updates should be reflected in volume [NodeConformance] [Conformance]","completed":346,"skipped":6462,"failed":0} +Feb 25 11:41:26.012: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "sched-pred-5830" for this suite. 02/25/23 11:41:26.018 +[AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:83 +{"msg":"PASSED [sig-scheduling] SchedulerPredicates [Serial] validates that NodeSelector is respected if not matching [Conformance]","completed":344,"skipped":6451,"failed":0} ------------------------------ -• [4.163 seconds] -[sig-storage] ConfigMap -test/e2e/common/storage/framework.go:23 - updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:123 +• [SLOW TEST] [7.375 seconds] +[sig-scheduling] SchedulerPredicates [Serial] +test/e2e/scheduling/framework.go:40 + validates that NodeSelector is respected if not matching [Conformance] + test/e2e/scheduling/predicates.go:438 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] ConfigMap + [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:09:12.585 - Dec 14 17:09:12.585: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename configmap 12/14/22 17:09:12.588 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:12.614 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:12.618 - [It] updates should be reflected in volume [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:123 - STEP: Creating configMap with name configmap-test-upd-1fdd77b9-36e8-475f-8116-2f796284a1fa 12/14/22 17:09:12.628 - STEP: Creating the pod 12/14/22 17:09:12.634 - Dec 14 17:09:12.645: INFO: Waiting up to 5m0s for pod "pod-configmaps-4c016e0a-9f52-4195-9639-37c11eb2e069" in namespace "configmap-9236" to be "running and ready" - Dec 14 17:09:12.651: INFO: Pod "pod-configmaps-4c016e0a-9f52-4195-9639-37c11eb2e069": Phase="Pending", Reason="", readiness=false. Elapsed: 6.180836ms - Dec 14 17:09:12.651: INFO: The phase of Pod pod-configmaps-4c016e0a-9f52-4195-9639-37c11eb2e069 is Pending, waiting for it to be Running (with Ready = true) - Dec 14 17:09:14.667: INFO: Pod "pod-configmaps-4c016e0a-9f52-4195-9639-37c11eb2e069": Phase="Running", Reason="", readiness=true. Elapsed: 2.021789128s - Dec 14 17:09:14.667: INFO: The phase of Pod pod-configmaps-4c016e0a-9f52-4195-9639-37c11eb2e069 is Running (Ready = true) - Dec 14 17:09:14.667: INFO: Pod "pod-configmaps-4c016e0a-9f52-4195-9639-37c11eb2e069" satisfied condition "running and ready" - STEP: Updating configmap configmap-test-upd-1fdd77b9-36e8-475f-8116-2f796284a1fa 12/14/22 17:09:14.696 - STEP: waiting to observe update in volume 12/14/22 17:09:14.71 - [AfterEach] [sig-storage] ConfigMap + STEP: Creating a kubernetes client 02/25/23 11:41:18.652 + Feb 25 11:41:18.653: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename sched-pred 02/25/23 11:41:18.66 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:18.722 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:18.727 + [BeforeEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:92 + Feb 25 11:41:18.731: INFO: Waiting up to 1m0s for all (but 0) nodes to be ready + Feb 25 11:41:18.749: INFO: Waiting for terminating namespaces to be deleted... + Feb 25 11:41:18.757: INFO: + Logging pods the apiserver thinks is on node diek3dua9gay-1 before test + Feb 25 11:41:18.776: INFO: echo-other-node-cd69fcf6b-hsnk9 from cilium-test started at 2023-02-25 09:38:06 +0000 UTC (2 container statuses recorded) + Feb 25 11:41:18.776: INFO: Container dns-test-server ready: true, restart count 0 + Feb 25 11:41:18.776: INFO: Container echo-other-node ready: true, restart count 0 + Feb 25 11:41:18.777: INFO: cilium-node-init-s84z5 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.777: INFO: Container node-init ready: true, restart count 0 + Feb 25 11:41:18.777: INFO: cilium-r4lc4 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.777: INFO: Container cilium-agent ready: true, restart count 0 + Feb 25 11:41:18.777: INFO: coredns-565d847f94-qqvf6 from kube-system started at 2023-02-25 08:49:06 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.777: INFO: Container coredns ready: true, restart count 0 + Feb 25 11:41:18.777: INFO: kube-addon-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.778: INFO: Container kube-addon-manager ready: true, restart count 0 + Feb 25 11:41:18.778: INFO: kube-apiserver-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:09 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.778: INFO: Container kube-apiserver ready: true, restart count 0 + Feb 25 11:41:18.778: INFO: kube-controller-manager-diek3dua9gay-1 from kube-system started at 2023-02-25 08:49:55 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.778: INFO: Container kube-controller-manager ready: true, restart count 0 + Feb 25 11:41:18.778: INFO: kube-proxy-cm4qg from kube-system started at 2023-02-25 08:46:38 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.778: INFO: Container kube-proxy ready: true, restart count 0 + Feb 25 11:41:18.778: INFO: kube-scheduler-diek3dua9gay-1 from kube-system started at 2023-02-25 08:46:27 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.778: INFO: Container kube-scheduler ready: true, restart count 0 + Feb 25 11:41:18.778: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-wznrx from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 11:41:18.778: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 11:41:18.778: INFO: Container systemd-logs ready: true, restart count 0 + Feb 25 11:41:18.778: INFO: + Logging pods the apiserver thinks is on node diek3dua9gay-2 before test + Feb 25 11:41:18.803: INFO: cilium-9tkfg from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.803: INFO: Container cilium-agent ready: true, restart count 0 + Feb 25 11:41:18.803: INFO: cilium-node-init-8b7n2 from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.803: INFO: Container node-init ready: true, restart count 0 + Feb 25 11:41:18.803: INFO: cilium-operator-59d7d77cdc-lppbd from kube-system started at 2023-02-25 08:47:47 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.803: INFO: Container cilium-operator ready: true, restart count 0 + Feb 25 11:41:18.803: INFO: coredns-565d847f94-5nh2n from kube-system started at 2023-02-25 08:48:51 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.803: INFO: Container coredns ready: true, restart count 0 + Feb 25 11:41:18.803: INFO: kube-addon-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:28 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.803: INFO: Container kube-addon-manager ready: true, restart count 0 + Feb 25 11:41:18.803: INFO: kube-apiserver-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.803: INFO: Container kube-apiserver ready: true, restart count 0 + Feb 25 11:41:18.803: INFO: kube-controller-manager-diek3dua9gay-2 from kube-system started at 2023-02-25 08:47:03 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.803: INFO: Container kube-controller-manager ready: true, restart count 0 + Feb 25 11:41:18.803: INFO: kube-proxy-d66m7 from kube-system started at 2023-02-25 08:46:57 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.803: INFO: Container kube-proxy ready: true, restart count 0 + Feb 25 11:41:18.803: INFO: kube-scheduler-diek3dua9gay-2 from kube-system started at 2023-02-25 08:50:15 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.803: INFO: Container kube-scheduler ready: true, restart count 0 + Feb 25 11:41:18.803: INFO: test-rs-946fw from replicaset-5096 started at 2023-02-25 11:41:18 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.803: INFO: Container httpd ready: false, restart count 0 + Feb 25 11:41:18.803: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-mml8w from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 11:41:18.803: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 11:41:18.803: INFO: Container systemd-logs ready: true, restart count 0 + Feb 25 11:41:18.803: INFO: + Logging pods the apiserver thinks is on node diek3dua9gay-3 before test + Feb 25 11:41:18.827: INFO: client-7b78db77d5-27xbt from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.827: INFO: Container client ready: true, restart count 0 + Feb 25 11:41:18.827: INFO: client2-78f748dd67-skwz2 from cilium-test started at 2023-02-25 09:57:06 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.827: INFO: Container client2 ready: true, restart count 0 + Feb 25 11:41:18.827: INFO: echo-same-node-85bc9b6b56-wqmfq from cilium-test started at 2023-02-25 09:57:46 +0000 UTC (2 container statuses recorded) + Feb 25 11:41:18.827: INFO: Container dns-test-server ready: true, restart count 0 + Feb 25 11:41:18.827: INFO: Container echo-same-node ready: true, restart count 0 + Feb 25 11:41:18.827: INFO: cilium-2cncz from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.828: INFO: Container cilium-agent ready: true, restart count 0 + Feb 25 11:41:18.828: INFO: cilium-node-init-979mb from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.828: INFO: Container node-init ready: true, restart count 0 + Feb 25 11:41:18.828: INFO: kube-proxy-22s9q from kube-system started at 2023-02-25 08:51:00 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.828: INFO: Container kube-proxy ready: true, restart count 0 + Feb 25 11:41:18.828: INFO: test-rs-2g7jr from replicaset-5096 started at 2023-02-25 11:41:13 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.828: INFO: Container httpd ready: true, restart count 0 + Feb 25 11:41:18.828: INFO: sonobuoy from sonobuoy started at 2023-02-25 09:45:14 +0000 UTC (1 container statuses recorded) + Feb 25 11:41:18.828: INFO: Container kube-sonobuoy ready: true, restart count 0 + Feb 25 11:41:18.828: INFO: sonobuoy-e2e-job-0b3c30975b5c4880 from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 11:41:18.828: INFO: Container e2e ready: true, restart count 0 + Feb 25 11:41:18.828: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 11:41:18.828: INFO: sonobuoy-systemd-logs-daemon-set-d97ff7768d5e483d-9pdjb from sonobuoy started at 2023-02-25 09:45:26 +0000 UTC (2 container statuses recorded) + Feb 25 11:41:18.828: INFO: Container sonobuoy-worker ready: true, restart count 0 + Feb 25 11:41:18.828: INFO: Container systemd-logs ready: true, restart count 0 + [It] validates that NodeSelector is respected if not matching [Conformance] + test/e2e/scheduling/predicates.go:438 + STEP: Trying to schedule Pod with nonempty NodeSelector. 02/25/23 11:41:18.829 + STEP: Considering event: + Type = [Warning], Name = [restricted-pod.17470e75b97347ce], Reason = [FailedScheduling], Message = [0/3 nodes are available: 3 node(s) didn't match Pod's node affinity/selector. preemption: 0/3 nodes are available: 3 Preemption is not helpful for scheduling.] 02/25/23 11:41:25.024 + [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] test/e2e/framework/framework.go:187 - Dec 14 17:09:16.734: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "configmap-9236" for this suite. 12/14/22 17:09:16.74 + Feb 25 11:41:26.012: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "sched-pred-5830" for this suite. 02/25/23 11:41:26.018 + [AfterEach] [sig-scheduling] SchedulerPredicates [Serial] + test/e2e/scheduling/predicates.go:83 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSSSS ------------------------------ -[sig-node] Container Runtime blackbox test on terminated container - should report termination message as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:231 -[BeforeEach] [sig-node] Container Runtime +[sig-node] Kubelet when scheduling an agnhost Pod with hostAliases + should write entries to /etc/hosts [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:148 +[BeforeEach] [sig-node] Kubelet test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:09:16.75 -Dec 14 17:09:16.750: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename container-runtime 12/14/22 17:09:16.752 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:16.776 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:16.78 -[It] should report termination message as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:231 -STEP: create the container 12/14/22 17:09:16.784 -STEP: wait for the container to reach Succeeded 12/14/22 17:09:16.8 -STEP: get the container status 12/14/22 17:09:19.832 -STEP: the container should be terminated 12/14/22 17:09:19.841 -STEP: the termination message should be set 12/14/22 17:09:19.842 -Dec 14 17:09:19.842: INFO: Expected: &{} to match Container's Termination Message: -- -STEP: delete the container 12/14/22 17:09:19.842 -[AfterEach] [sig-node] Container Runtime +STEP: Creating a kubernetes client 02/25/23 11:41:26.028 +Feb 25 11:41:26.028: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename kubelet-test 02/25/23 11:41:26.033 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:26.056 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:26.06 +[BeforeEach] [sig-node] Kubelet + test/e2e/common/node/kubelet.go:41 +[It] should write entries to /etc/hosts [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:148 +STEP: Waiting for pod completion 02/25/23 11:41:26.077 +Feb 25 11:41:26.077: INFO: Waiting up to 3m0s for pod "agnhost-host-aliasesd8401727-076e-48af-a2b0-d44fb0ec4b59" in namespace "kubelet-test-1515" to be "completed" +Feb 25 11:41:26.082: INFO: Pod "agnhost-host-aliasesd8401727-076e-48af-a2b0-d44fb0ec4b59": Phase="Pending", Reason="", readiness=false. Elapsed: 4.486351ms +Feb 25 11:41:28.090: INFO: Pod "agnhost-host-aliasesd8401727-076e-48af-a2b0-d44fb0ec4b59": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012697784s +Feb 25 11:41:30.089: INFO: Pod "agnhost-host-aliasesd8401727-076e-48af-a2b0-d44fb0ec4b59": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011570946s +Feb 25 11:41:32.090: INFO: Pod "agnhost-host-aliasesd8401727-076e-48af-a2b0-d44fb0ec4b59": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.01265626s +Feb 25 11:41:32.090: INFO: Pod "agnhost-host-aliasesd8401727-076e-48af-a2b0-d44fb0ec4b59" satisfied condition "completed" +[AfterEach] [sig-node] Kubelet test/e2e/framework/framework.go:187 -Dec 14 17:09:19.871: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "container-runtime-4994" for this suite. 12/14/22 17:09:19.877 -{"msg":"PASSED [sig-node] Container Runtime blackbox test on terminated container should report termination message as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance]","completed":347,"skipped":6465,"failed":0} +Feb 25 11:41:32.179: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "kubelet-test-1515" for this suite. 02/25/23 11:41:32.186 +{"msg":"PASSED [sig-node] Kubelet when scheduling an agnhost Pod with hostAliases should write entries to /etc/hosts [NodeConformance] [Conformance]","completed":345,"skipped":6459,"failed":0} ------------------------------ -• [3.134 seconds] -[sig-node] Container Runtime +• [SLOW TEST] [6.181 seconds] +[sig-node] Kubelet test/e2e/common/node/framework.go:23 - blackbox test - test/e2e/common/node/runtime.go:43 - on terminated container - test/e2e/common/node/runtime.go:136 - should report termination message as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:231 + when scheduling an agnhost Pod with hostAliases + test/e2e/common/node/kubelet.go:140 + should write entries to /etc/hosts [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:148 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-node] Container Runtime + [BeforeEach] [sig-node] Kubelet test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:09:16.75 - Dec 14 17:09:16.750: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename container-runtime 12/14/22 17:09:16.752 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:16.776 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:16.78 - [It] should report termination message as empty when pod succeeds and TerminationMessagePolicy FallbackToLogsOnError is set [NodeConformance] [Conformance] - test/e2e/common/node/runtime.go:231 - STEP: create the container 12/14/22 17:09:16.784 - STEP: wait for the container to reach Succeeded 12/14/22 17:09:16.8 - STEP: get the container status 12/14/22 17:09:19.832 - STEP: the container should be terminated 12/14/22 17:09:19.841 - STEP: the termination message should be set 12/14/22 17:09:19.842 - Dec 14 17:09:19.842: INFO: Expected: &{} to match Container's Termination Message: -- - STEP: delete the container 12/14/22 17:09:19.842 - [AfterEach] [sig-node] Container Runtime + STEP: Creating a kubernetes client 02/25/23 11:41:26.028 + Feb 25 11:41:26.028: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename kubelet-test 02/25/23 11:41:26.033 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:26.056 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:26.06 + [BeforeEach] [sig-node] Kubelet + test/e2e/common/node/kubelet.go:41 + [It] should write entries to /etc/hosts [NodeConformance] [Conformance] + test/e2e/common/node/kubelet.go:148 + STEP: Waiting for pod completion 02/25/23 11:41:26.077 + Feb 25 11:41:26.077: INFO: Waiting up to 3m0s for pod "agnhost-host-aliasesd8401727-076e-48af-a2b0-d44fb0ec4b59" in namespace "kubelet-test-1515" to be "completed" + Feb 25 11:41:26.082: INFO: Pod "agnhost-host-aliasesd8401727-076e-48af-a2b0-d44fb0ec4b59": Phase="Pending", Reason="", readiness=false. Elapsed: 4.486351ms + Feb 25 11:41:28.090: INFO: Pod "agnhost-host-aliasesd8401727-076e-48af-a2b0-d44fb0ec4b59": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012697784s + Feb 25 11:41:30.089: INFO: Pod "agnhost-host-aliasesd8401727-076e-48af-a2b0-d44fb0ec4b59": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011570946s + Feb 25 11:41:32.090: INFO: Pod "agnhost-host-aliasesd8401727-076e-48af-a2b0-d44fb0ec4b59": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.01265626s + Feb 25 11:41:32.090: INFO: Pod "agnhost-host-aliasesd8401727-076e-48af-a2b0-d44fb0ec4b59" satisfied condition "completed" + [AfterEach] [sig-node] Kubelet test/e2e/framework/framework.go:187 - Dec 14 17:09:19.871: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "container-runtime-4994" for this suite. 12/14/22 17:09:19.877 + Feb 25 11:41:32.179: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "kubelet-test-1515" for this suite. 02/25/23 11:41:32.186 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSS +SSSS ------------------------------ -[sig-api-machinery] ResourceQuota - should be able to update and delete ResourceQuota. [Conformance] - test/e2e/apimachinery/resource_quota.go:874 -[BeforeEach] [sig-api-machinery] ResourceQuota +[sig-network] EndpointSlice + should support creating EndpointSlice API operations [Conformance] + test/e2e/network/endpointslice.go:352 +[BeforeEach] [sig-network] EndpointSlice test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:09:19.887 -Dec 14 17:09:19.888: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename resourcequota 12/14/22 17:09:19.891 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:19.91 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:19.913 -[It] should be able to update and delete ResourceQuota. [Conformance] - test/e2e/apimachinery/resource_quota.go:874 -STEP: Creating a ResourceQuota 12/14/22 17:09:19.916 -STEP: Getting a ResourceQuota 12/14/22 17:09:19.923 -STEP: Updating a ResourceQuota 12/14/22 17:09:19.931 -STEP: Verifying a ResourceQuota was modified 12/14/22 17:09:19.986 -STEP: Deleting a ResourceQuota 12/14/22 17:09:19.993 -STEP: Verifying the deleted ResourceQuota 12/14/22 17:09:20.016 -[AfterEach] [sig-api-machinery] ResourceQuota +STEP: Creating a kubernetes client 02/25/23 11:41:32.212 +Feb 25 11:41:32.212: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename endpointslice 02/25/23 11:41:32.215 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:32.244 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:32.252 +[BeforeEach] [sig-network] EndpointSlice + test/e2e/network/endpointslice.go:51 +[It] should support creating EndpointSlice API operations [Conformance] + test/e2e/network/endpointslice.go:352 +STEP: getting /apis 02/25/23 11:41:32.256 +STEP: getting /apis/discovery.k8s.io 02/25/23 11:41:32.259 +STEP: getting /apis/discovery.k8s.iov1 02/25/23 11:41:32.261 +STEP: creating 02/25/23 11:41:32.264 +STEP: getting 02/25/23 11:41:32.287 +STEP: listing 02/25/23 11:41:32.299 +STEP: watching 02/25/23 11:41:32.305 +Feb 25 11:41:32.305: INFO: starting watch +STEP: cluster-wide listing 02/25/23 11:41:32.308 +STEP: cluster-wide watching 02/25/23 11:41:32.314 +Feb 25 11:41:32.314: INFO: starting watch +STEP: patching 02/25/23 11:41:32.316 +STEP: updating 02/25/23 11:41:32.326 +Feb 25 11:41:32.338: INFO: waiting for watch events with expected annotations +Feb 25 11:41:32.338: INFO: saw patched and updated annotations +STEP: deleting 02/25/23 11:41:32.338 +STEP: deleting a collection 02/25/23 11:41:32.358 +[AfterEach] [sig-network] EndpointSlice test/e2e/framework/framework.go:187 -Dec 14 17:09:20.020: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "resourcequota-4630" for this suite. 12/14/22 17:09:20.025 -{"msg":"PASSED [sig-api-machinery] ResourceQuota should be able to update and delete ResourceQuota. [Conformance]","completed":348,"skipped":6473,"failed":0} +Feb 25 11:41:32.383: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "endpointslice-4980" for this suite. 02/25/23 11:41:32.391 +{"msg":"PASSED [sig-network] EndpointSlice should support creating EndpointSlice API operations [Conformance]","completed":346,"skipped":6463,"failed":0} ------------------------------ -• [0.144 seconds] -[sig-api-machinery] ResourceQuota -test/e2e/apimachinery/framework.go:23 - should be able to update and delete ResourceQuota. [Conformance] - test/e2e/apimachinery/resource_quota.go:874 +• [0.190 seconds] +[sig-network] EndpointSlice +test/e2e/network/common/framework.go:23 + should support creating EndpointSlice API operations [Conformance] + test/e2e/network/endpointslice.go:352 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] ResourceQuota + [BeforeEach] [sig-network] EndpointSlice test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:09:19.887 - Dec 14 17:09:19.888: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename resourcequota 12/14/22 17:09:19.891 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:19.91 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:19.913 - [It] should be able to update and delete ResourceQuota. [Conformance] - test/e2e/apimachinery/resource_quota.go:874 - STEP: Creating a ResourceQuota 12/14/22 17:09:19.916 - STEP: Getting a ResourceQuota 12/14/22 17:09:19.923 - STEP: Updating a ResourceQuota 12/14/22 17:09:19.931 - STEP: Verifying a ResourceQuota was modified 12/14/22 17:09:19.986 - STEP: Deleting a ResourceQuota 12/14/22 17:09:19.993 - STEP: Verifying the deleted ResourceQuota 12/14/22 17:09:20.016 - [AfterEach] [sig-api-machinery] ResourceQuota + STEP: Creating a kubernetes client 02/25/23 11:41:32.212 + Feb 25 11:41:32.212: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename endpointslice 02/25/23 11:41:32.215 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:32.244 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:32.252 + [BeforeEach] [sig-network] EndpointSlice + test/e2e/network/endpointslice.go:51 + [It] should support creating EndpointSlice API operations [Conformance] + test/e2e/network/endpointslice.go:352 + STEP: getting /apis 02/25/23 11:41:32.256 + STEP: getting /apis/discovery.k8s.io 02/25/23 11:41:32.259 + STEP: getting /apis/discovery.k8s.iov1 02/25/23 11:41:32.261 + STEP: creating 02/25/23 11:41:32.264 + STEP: getting 02/25/23 11:41:32.287 + STEP: listing 02/25/23 11:41:32.299 + STEP: watching 02/25/23 11:41:32.305 + Feb 25 11:41:32.305: INFO: starting watch + STEP: cluster-wide listing 02/25/23 11:41:32.308 + STEP: cluster-wide watching 02/25/23 11:41:32.314 + Feb 25 11:41:32.314: INFO: starting watch + STEP: patching 02/25/23 11:41:32.316 + STEP: updating 02/25/23 11:41:32.326 + Feb 25 11:41:32.338: INFO: waiting for watch events with expected annotations + Feb 25 11:41:32.338: INFO: saw patched and updated annotations + STEP: deleting 02/25/23 11:41:32.338 + STEP: deleting a collection 02/25/23 11:41:32.358 + [AfterEach] [sig-network] EndpointSlice test/e2e/framework/framework.go:187 - Dec 14 17:09:20.020: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "resourcequota-4630" for this suite. 12/14/22 17:09:20.025 + Feb 25 11:41:32.383: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "endpointslice-4980" for this suite. 02/25/23 11:41:32.391 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSS ------------------------------ -[sig-apps] Deployment - Deployment should have a working scale subresource [Conformance] - test/e2e/apps/deployment.go:150 -[BeforeEach] [sig-apps] Deployment +[sig-node] InitContainer [NodeConformance] + should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] + test/e2e/common/node/init_container.go:457 +[BeforeEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:09:20.054 -Dec 14 17:09:20.055: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename deployment 12/14/22 17:09:20.058 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:20.09 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:20.095 -[BeforeEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:91 -[It] Deployment should have a working scale subresource [Conformance] - test/e2e/apps/deployment.go:150 -Dec 14 17:09:20.099: INFO: Creating simple deployment test-new-deployment -Dec 14 17:09:20.128: INFO: deployment "test-new-deployment" doesn't have the required revision set -Dec 14 17:09:22.152: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 20, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 20, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 20, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 20, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-new-deployment-845c8977d9\" is progressing."}}, CollisionCount:(*int32)(nil)} -STEP: getting scale subresource 12/14/22 17:09:24.165 -STEP: updating a scale subresource 12/14/22 17:09:24.172 -STEP: verifying the deployment Spec.Replicas was modified 12/14/22 17:09:24.185 -STEP: Patch a scale subresource 12/14/22 17:09:24.191 -[AfterEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:84 -Dec 14 17:09:24.239: INFO: Deployment "test-new-deployment": -&Deployment{ObjectMeta:{test-new-deployment deployment-4129 bb3aa406-7635-4ad4-9dc6-ae449854e2db 31261 3 2022-12-14 17:09:20 +0000 UTC map[name:httpd] map[deployment.kubernetes.io/revision:1] [] [] [{e2e.test Update apps/v1 FieldsV1 {"f:spec":{"f:replicas":{}}} scale} {e2e.test Update apps/v1 2022-12-14 17:09:20 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 17:09:22 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*4,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0035e4a28 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2022-12-14 17:09:22 +0000 UTC,LastTransitionTime:2022-12-14 17:09:22 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-new-deployment-845c8977d9" has successfully progressed.,LastUpdateTime:2022-12-14 17:09:22 +0000 UTC,LastTransitionTime:2022-12-14 17:09:20 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} - -Dec 14 17:09:24.258: INFO: New ReplicaSet "test-new-deployment-845c8977d9" of Deployment "test-new-deployment": -&ReplicaSet{ObjectMeta:{test-new-deployment-845c8977d9 deployment-4129 05806ecd-5a8c-440e-a412-41760360c72d 31265 2 2022-12-14 17:09:20 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[deployment.kubernetes.io/desired-replicas:2 deployment.kubernetes.io/max-replicas:3 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-new-deployment bb3aa406-7635-4ad4-9dc6-ae449854e2db 0xc0035e4e77 0xc0035e4e78}] [] [{kube-controller-manager Update apps/v1 2022-12-14 17:09:24 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"bb3aa406-7635-4ad4-9dc6-ae449854e2db\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 17:09:24 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*2,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 845c8977d9,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0035e4f08 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} -Dec 14 17:09:24.265: INFO: Pod "test-new-deployment-845c8977d9-2tcq2" is not available: -&Pod{ObjectMeta:{test-new-deployment-845c8977d9-2tcq2 test-new-deployment-845c8977d9- deployment-4129 c288a84b-9678-4a6d-8604-1eee97a355af 31269 0 2022-12-14 17:09:24 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet test-new-deployment-845c8977d9 05806ecd-5a8c-440e-a412-41760360c72d 0xc0035e52b7 0xc0035e52b8}] [] [{kube-controller-manager Update v1 2022-12-14 17:09:24 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"05806ecd-5a8c-440e-a412-41760360c72d\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 17:09:24 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-rwnw5,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rwnw5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:24 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:24 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:24 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:24 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 17:09:24 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} -Dec 14 17:09:24.266: INFO: Pod "test-new-deployment-845c8977d9-6xcn7" is available: -&Pod{ObjectMeta:{test-new-deployment-845c8977d9-6xcn7 test-new-deployment-845c8977d9- deployment-4129 087c6345-bcd6-4026-929c-a6987e50a382 31252 0 2022-12-14 17:09:20 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet test-new-deployment-845c8977d9 05806ecd-5a8c-440e-a412-41760360c72d 0xc0035e5487 0xc0035e5488}] [] [{kube-controller-manager Update v1 2022-12-14 17:09:20 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"05806ecd-5a8c-440e-a412-41760360c72d\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 17:09:22 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.201\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-nfrfs,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nfrfs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:20 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:22 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:22 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:20 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.201,StartTime:2022-12-14 17:09:20 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 17:09:21 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://9c6e39d223342eafb5dbf8257b0ce77969dbc78a3254621b8695e04a65e508a7,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.201,},},EphemeralContainerStatuses:[]ContainerStatus{},},} -[AfterEach] [sig-apps] Deployment +STEP: Creating a kubernetes client 02/25/23 11:41:32.403 +Feb 25 11:41:32.403: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename init-container 02/25/23 11:41:32.405 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:32.436 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:32.441 +[BeforeEach] [sig-node] InitContainer [NodeConformance] + test/e2e/common/node/init_container.go:164 +[It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] + test/e2e/common/node/init_container.go:457 +STEP: creating the pod 02/25/23 11:41:32.444 +Feb 25 11:41:32.445: INFO: PodSpec: initContainers in spec.initContainers +[AfterEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:187 -Dec 14 17:09:24.266: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "deployment-4129" for this suite. 12/14/22 17:09:24.273 -{"msg":"PASSED [sig-apps] Deployment Deployment should have a working scale subresource [Conformance]","completed":349,"skipped":6495,"failed":0} +Feb 25 11:41:37.364: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "init-container-5991" for this suite. 02/25/23 11:41:37.375 +{"msg":"PASSED [sig-node] InitContainer [NodeConformance] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance]","completed":347,"skipped":6472,"failed":0} ------------------------------ -• [4.231 seconds] -[sig-apps] Deployment -test/e2e/apps/framework.go:23 - Deployment should have a working scale subresource [Conformance] - test/e2e/apps/deployment.go:150 +• [4.991 seconds] +[sig-node] InitContainer [NodeConformance] +test/e2e/common/node/framework.go:23 + should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] + test/e2e/common/node/init_container.go:457 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] Deployment + [BeforeEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:09:20.054 - Dec 14 17:09:20.055: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename deployment 12/14/22 17:09:20.058 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:20.09 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:20.095 - [BeforeEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:91 - [It] Deployment should have a working scale subresource [Conformance] - test/e2e/apps/deployment.go:150 - Dec 14 17:09:20.099: INFO: Creating simple deployment test-new-deployment - Dec 14 17:09:20.128: INFO: deployment "test-new-deployment" doesn't have the required revision set - Dec 14 17:09:22.152: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 20, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 20, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 20, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 20, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-new-deployment-845c8977d9\" is progressing."}}, CollisionCount:(*int32)(nil)} - STEP: getting scale subresource 12/14/22 17:09:24.165 - STEP: updating a scale subresource 12/14/22 17:09:24.172 - STEP: verifying the deployment Spec.Replicas was modified 12/14/22 17:09:24.185 - STEP: Patch a scale subresource 12/14/22 17:09:24.191 - [AfterEach] [sig-apps] Deployment - test/e2e/apps/deployment.go:84 - Dec 14 17:09:24.239: INFO: Deployment "test-new-deployment": - &Deployment{ObjectMeta:{test-new-deployment deployment-4129 bb3aa406-7635-4ad4-9dc6-ae449854e2db 31261 3 2022-12-14 17:09:20 +0000 UTC map[name:httpd] map[deployment.kubernetes.io/revision:1] [] [] [{e2e.test Update apps/v1 FieldsV1 {"f:spec":{"f:replicas":{}}} scale} {e2e.test Update apps/v1 2022-12-14 17:09:20 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 17:09:22 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*4,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0035e4a28 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:1,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2022-12-14 17:09:22 +0000 UTC,LastTransitionTime:2022-12-14 17:09:22 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-new-deployment-845c8977d9" has successfully progressed.,LastUpdateTime:2022-12-14 17:09:22 +0000 UTC,LastTransitionTime:2022-12-14 17:09:20 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} - - Dec 14 17:09:24.258: INFO: New ReplicaSet "test-new-deployment-845c8977d9" of Deployment "test-new-deployment": - &ReplicaSet{ObjectMeta:{test-new-deployment-845c8977d9 deployment-4129 05806ecd-5a8c-440e-a412-41760360c72d 31265 2 2022-12-14 17:09:20 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[deployment.kubernetes.io/desired-replicas:2 deployment.kubernetes.io/max-replicas:3 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-new-deployment bb3aa406-7635-4ad4-9dc6-ae449854e2db 0xc0035e4e77 0xc0035e4e78}] [] [{kube-controller-manager Update apps/v1 2022-12-14 17:09:24 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"bb3aa406-7635-4ad4-9dc6-ae449854e2db\"}":{}}},"f:spec":{"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 17:09:24 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*2,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: httpd,pod-template-hash: 845c8977d9,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0035e4f08 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} - Dec 14 17:09:24.265: INFO: Pod "test-new-deployment-845c8977d9-2tcq2" is not available: - &Pod{ObjectMeta:{test-new-deployment-845c8977d9-2tcq2 test-new-deployment-845c8977d9- deployment-4129 c288a84b-9678-4a6d-8604-1eee97a355af 31269 0 2022-12-14 17:09:24 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet test-new-deployment-845c8977d9 05806ecd-5a8c-440e-a412-41760360c72d 0xc0035e52b7 0xc0035e52b8}] [] [{kube-controller-manager Update v1 2022-12-14 17:09:24 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"05806ecd-5a8c-440e-a412-41760360c72d\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 17:09:24 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-rwnw5,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rwnw5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-2,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Pending,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:24 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:24 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:ContainersReady,Status:False,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:24 +0000 UTC,Reason:ContainersNotReady,Message:containers with unready status: [httpd],},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:24 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.197,PodIP:,StartTime:2022-12-14 17:09:24 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:&ContainerStateWaiting{Reason:ContainerCreating,Message:,},Running:nil,Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:false,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:,ContainerID:,Started:*false,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{},EphemeralContainerStatuses:[]ContainerStatus{},},} - Dec 14 17:09:24.266: INFO: Pod "test-new-deployment-845c8977d9-6xcn7" is available: - &Pod{ObjectMeta:{test-new-deployment-845c8977d9-6xcn7 test-new-deployment-845c8977d9- deployment-4129 087c6345-bcd6-4026-929c-a6987e50a382 31252 0 2022-12-14 17:09:20 +0000 UTC map[name:httpd pod-template-hash:845c8977d9] map[] [{apps/v1 ReplicaSet test-new-deployment-845c8977d9 05806ecd-5a8c-440e-a412-41760360c72d 0xc0035e5487 0xc0035e5488}] [] [{kube-controller-manager Update v1 2022-12-14 17:09:20 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"05806ecd-5a8c-440e-a412-41760360c72d\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 17:09:22 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.201\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-nfrfs,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:httpd,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nfrfs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:20 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:22 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:22 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:20 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.201,StartTime:2022-12-14 17:09:20 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:httpd,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 17:09:21 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/httpd:2.4.38-2,ImageID:registry.k8s.io/e2e-test-images/httpd@sha256:1b9d1b2f36cb2dbee1960e82a9344aeb11bd4c4c03abf5e1853e0559c23855e3,ContainerID:cri-o://9c6e39d223342eafb5dbf8257b0ce77969dbc78a3254621b8695e04a65e508a7,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.201,},},EphemeralContainerStatuses:[]ContainerStatus{},},} - [AfterEach] [sig-apps] Deployment + STEP: Creating a kubernetes client 02/25/23 11:41:32.403 + Feb 25 11:41:32.403: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename init-container 02/25/23 11:41:32.405 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:32.436 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:32.441 + [BeforeEach] [sig-node] InitContainer [NodeConformance] + test/e2e/common/node/init_container.go:164 + [It] should not start app containers and fail the pod if init containers fail on a RestartNever pod [Conformance] + test/e2e/common/node/init_container.go:457 + STEP: creating the pod 02/25/23 11:41:32.444 + Feb 25 11:41:32.445: INFO: PodSpec: initContainers in spec.initContainers + [AfterEach] [sig-node] InitContainer [NodeConformance] test/e2e/framework/framework.go:187 - Dec 14 17:09:24.266: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "deployment-4129" for this suite. 12/14/22 17:09:24.273 + Feb 25 11:41:37.364: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "init-container-5991" for this suite. 02/25/23 11:41:37.375 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSS +SSSSS ------------------------------ -[sig-storage] Projected downwardAPI - should provide container's cpu limit [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:192 -[BeforeEach] [sig-storage] Projected downwardAPI +[sig-api-machinery] Watchers + should observe add, update, and delete watch notifications on configmaps [Conformance] + test/e2e/apimachinery/watch.go:60 +[BeforeEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:09:24.291 -Dec 14 17:09:24.291: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename projected 12/14/22 17:09:24.295 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:24.353 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:24.356 -[BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 -[It] should provide container's cpu limit [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:192 -STEP: Creating a pod to test downward API volume plugin 12/14/22 17:09:24.36 -Dec 14 17:09:24.374: INFO: Waiting up to 5m0s for pod "downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c" in namespace "projected-3948" to be "Succeeded or Failed" -Dec 14 17:09:24.378: INFO: Pod "downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.486049ms -Dec 14 17:09:26.384: INFO: Pod "downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010286392s -Dec 14 17:09:28.385: INFO: Pod "downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011421245s -Dec 14 17:09:30.386: INFO: Pod "downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.012275994s -STEP: Saw pod success 12/14/22 17:09:30.386 -Dec 14 17:09:30.386: INFO: Pod "downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c" satisfied condition "Succeeded or Failed" -Dec 14 17:09:30.393: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c container client-container: -STEP: delete the pod 12/14/22 17:09:30.402 -Dec 14 17:09:30.420: INFO: Waiting for pod downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c to disappear -Dec 14 17:09:30.425: INFO: Pod downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c no longer exists -[AfterEach] [sig-storage] Projected downwardAPI +STEP: Creating a kubernetes client 02/25/23 11:41:37.399 +Feb 25 11:41:37.399: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename watch 02/25/23 11:41:37.405 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:37.437 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:37.441 +[It] should observe add, update, and delete watch notifications on configmaps [Conformance] + test/e2e/apimachinery/watch.go:60 +STEP: creating a watch on configmaps with label A 02/25/23 11:41:37.446 +STEP: creating a watch on configmaps with label B 02/25/23 11:41:37.448 +STEP: creating a watch on configmaps with label A or B 02/25/23 11:41:37.451 +STEP: creating a configmap with label A and ensuring the correct watchers observe the notification 02/25/23 11:41:37.453 +Feb 25 11:41:37.462: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47132 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} +Feb 25 11:41:37.463: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47132 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} +STEP: modifying configmap A and ensuring the correct watchers observe the notification 02/25/23 11:41:37.463 +Feb 25 11:41:37.479: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47133 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} +Feb 25 11:41:37.479: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47133 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} +STEP: modifying configmap A again and ensuring the correct watchers observe the notification 02/25/23 11:41:37.48 +Feb 25 11:41:37.497: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47135 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} +Feb 25 11:41:37.497: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47135 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} +STEP: deleting configmap A and ensuring the correct watchers observe the notification 02/25/23 11:41:37.497 +Feb 25 11:41:37.511: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47136 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} +Feb 25 11:41:37.512: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47136 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} +STEP: creating a configmap with label B and ensuring the correct watchers observe the notification 02/25/23 11:41:37.512 +Feb 25 11:41:37.523: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-9770 9ac9e0e9-26c4-4254-9c1a-4d1fe48944ec 47137 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} +Feb 25 11:41:37.523: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-9770 9ac9e0e9-26c4-4254-9c1a-4d1fe48944ec 47137 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} +STEP: deleting configmap B and ensuring the correct watchers observe the notification 02/25/23 11:41:47.524 +Feb 25 11:41:47.536: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-9770 9ac9e0e9-26c4-4254-9c1a-4d1fe48944ec 47181 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} +Feb 25 11:41:47.536: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-9770 9ac9e0e9-26c4-4254-9c1a-4d1fe48944ec 47181 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} +[AfterEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:187 -Dec 14 17:09:30.425: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "projected-3948" for this suite. 12/14/22 17:09:30.433 -{"msg":"PASSED [sig-storage] Projected downwardAPI should provide container's cpu limit [NodeConformance] [Conformance]","completed":350,"skipped":6505,"failed":0} +Feb 25 11:41:57.538: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "watch-9770" for this suite. 02/25/23 11:41:57.546 +{"msg":"PASSED [sig-api-machinery] Watchers should observe add, update, and delete watch notifications on configmaps [Conformance]","completed":348,"skipped":6477,"failed":0} ------------------------------ -• [SLOW TEST] [6.151 seconds] -[sig-storage] Projected downwardAPI -test/e2e/common/storage/framework.go:23 - should provide container's cpu limit [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:192 +• [SLOW TEST] [20.170 seconds] +[sig-api-machinery] Watchers +test/e2e/apimachinery/framework.go:23 + should observe add, update, and delete watch notifications on configmaps [Conformance] + test/e2e/apimachinery/watch.go:60 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Projected downwardAPI + [BeforeEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:09:24.291 - Dec 14 17:09:24.291: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename projected 12/14/22 17:09:24.295 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:24.353 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:24.356 - [BeforeEach] [sig-storage] Projected downwardAPI - test/e2e/common/storage/projected_downwardapi.go:43 - [It] should provide container's cpu limit [NodeConformance] [Conformance] - test/e2e/common/storage/projected_downwardapi.go:192 - STEP: Creating a pod to test downward API volume plugin 12/14/22 17:09:24.36 - Dec 14 17:09:24.374: INFO: Waiting up to 5m0s for pod "downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c" in namespace "projected-3948" to be "Succeeded or Failed" - Dec 14 17:09:24.378: INFO: Pod "downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.486049ms - Dec 14 17:09:26.384: INFO: Pod "downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c": Phase="Pending", Reason="", readiness=false. Elapsed: 2.010286392s - Dec 14 17:09:28.385: INFO: Pod "downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c": Phase="Pending", Reason="", readiness=false. Elapsed: 4.011421245s - Dec 14 17:09:30.386: INFO: Pod "downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.012275994s - STEP: Saw pod success 12/14/22 17:09:30.386 - Dec 14 17:09:30.386: INFO: Pod "downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c" satisfied condition "Succeeded or Failed" - Dec 14 17:09:30.393: INFO: Trying to get logs from node kiedai4iakae-3 pod downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c container client-container: - STEP: delete the pod 12/14/22 17:09:30.402 - Dec 14 17:09:30.420: INFO: Waiting for pod downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c to disappear - Dec 14 17:09:30.425: INFO: Pod downwardapi-volume-2dc1886c-3363-4c30-9b8a-285a6dab279c no longer exists - [AfterEach] [sig-storage] Projected downwardAPI + STEP: Creating a kubernetes client 02/25/23 11:41:37.399 + Feb 25 11:41:37.399: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename watch 02/25/23 11:41:37.405 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:37.437 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:37.441 + [It] should observe add, update, and delete watch notifications on configmaps [Conformance] + test/e2e/apimachinery/watch.go:60 + STEP: creating a watch on configmaps with label A 02/25/23 11:41:37.446 + STEP: creating a watch on configmaps with label B 02/25/23 11:41:37.448 + STEP: creating a watch on configmaps with label A or B 02/25/23 11:41:37.451 + STEP: creating a configmap with label A and ensuring the correct watchers observe the notification 02/25/23 11:41:37.453 + Feb 25 11:41:37.462: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47132 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} + Feb 25 11:41:37.463: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47132 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} + STEP: modifying configmap A and ensuring the correct watchers observe the notification 02/25/23 11:41:37.463 + Feb 25 11:41:37.479: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47133 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} + Feb 25 11:41:37.479: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47133 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 1,},BinaryData:map[string][]byte{},Immutable:nil,} + STEP: modifying configmap A again and ensuring the correct watchers observe the notification 02/25/23 11:41:37.48 + Feb 25 11:41:37.497: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47135 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} + Feb 25 11:41:37.497: INFO: Got : MODIFIED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47135 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} + STEP: deleting configmap A and ensuring the correct watchers observe the notification 02/25/23 11:41:37.497 + Feb 25 11:41:37.511: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47136 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} + Feb 25 11:41:37.512: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-a watch-9770 87979e0e-bca6-40f0-806a-bbb51da0959a 47136 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-A] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:data":{".":{},"f:mutation":{}},"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{mutation: 2,},BinaryData:map[string][]byte{},Immutable:nil,} + STEP: creating a configmap with label B and ensuring the correct watchers observe the notification 02/25/23 11:41:37.512 + Feb 25 11:41:37.523: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-9770 9ac9e0e9-26c4-4254-9c1a-4d1fe48944ec 47137 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} + Feb 25 11:41:37.523: INFO: Got : ADDED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-9770 9ac9e0e9-26c4-4254-9c1a-4d1fe48944ec 47137 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} + STEP: deleting configmap B and ensuring the correct watchers observe the notification 02/25/23 11:41:47.524 + Feb 25 11:41:47.536: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-9770 9ac9e0e9-26c4-4254-9c1a-4d1fe48944ec 47181 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} + Feb 25 11:41:47.536: INFO: Got : DELETED &ConfigMap{ObjectMeta:{e2e-watch-test-configmap-b watch-9770 9ac9e0e9-26c4-4254-9c1a-4d1fe48944ec 47181 0 2023-02-25 11:41:37 +0000 UTC map[watch-this-configmap:multiple-watchers-B] map[] [] [] [{e2e.test Update v1 2023-02-25 11:41:37 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:watch-this-configmap":{}}}} }]},Data:map[string]string{},BinaryData:map[string][]byte{},Immutable:nil,} + [AfterEach] [sig-api-machinery] Watchers test/e2e/framework/framework.go:187 - Dec 14 17:09:30.425: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "projected-3948" for this suite. 12/14/22 17:09:30.433 + Feb 25 11:41:57.538: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "watch-9770" for this suite. 02/25/23 11:41:57.546 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-apps] Deployment - deployment should support rollover [Conformance] - test/e2e/apps/deployment.go:132 + deployment should delete old replica sets [Conformance] + test/e2e/apps/deployment.go:122 [BeforeEach] [sig-apps] Deployment test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:09:30.451 -Dec 14 17:09:30.451: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename deployment 12/14/22 17:09:30.452 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:30.491 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:30.494 +STEP: Creating a kubernetes client 02/25/23 11:41:57.578 +Feb 25 11:41:57.579: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename deployment 02/25/23 11:41:57.581 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:57.607 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:57.611 [BeforeEach] [sig-apps] Deployment test/e2e/apps/deployment.go:91 -[It] deployment should support rollover [Conformance] - test/e2e/apps/deployment.go:132 -Dec 14 17:09:30.527: INFO: Pod name rollover-pod: Found 0 pods out of 1 -Dec 14 17:09:35.532: INFO: Pod name rollover-pod: Found 1 pods out of 1 -STEP: ensuring each pod is running 12/14/22 17:09:35.532 -Dec 14 17:09:35.533: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready -Dec 14 17:09:37.543: INFO: Creating deployment "test-rollover-deployment" -Dec 14 17:09:37.560: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations -Dec 14 17:09:39.574: INFO: Check revision of new replica set for deployment "test-rollover-deployment" -Dec 14 17:09:39.584: INFO: Ensure that both replica sets have 1 created replica -Dec 14 17:09:39.593: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update -Dec 14 17:09:39.610: INFO: Updating deployment test-rollover-deployment -Dec 14 17:09:39.610: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller -Dec 14 17:09:41.622: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2 -Dec 14 17:09:41.634: INFO: Make sure deployment "test-rollover-deployment" is complete -Dec 14 17:09:41.644: INFO: all replica sets need to contain the pod-template-hash label -Dec 14 17:09:41.644: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 41, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 17:09:43.658: INFO: all replica sets need to contain the pod-template-hash label -Dec 14 17:09:43.658: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 41, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 17:09:45.656: INFO: all replica sets need to contain the pod-template-hash label -Dec 14 17:09:45.656: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 41, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 17:09:47.658: INFO: all replica sets need to contain the pod-template-hash label -Dec 14 17:09:47.658: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 41, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 17:09:49.658: INFO: all replica sets need to contain the pod-template-hash label -Dec 14 17:09:49.659: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 41, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 17:09:51.685: INFO: -Dec 14 17:09:51.685: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:2, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 51, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} -Dec 14 17:09:53.659: INFO: -Dec 14 17:09:53.659: INFO: Ensure that both old replica sets have no replicas +[It] deployment should delete old replica sets [Conformance] + test/e2e/apps/deployment.go:122 +Feb 25 11:41:57.645: INFO: Pod name cleanup-pod: Found 0 pods out of 1 +Feb 25 11:42:02.652: INFO: Pod name cleanup-pod: Found 1 pods out of 1 +STEP: ensuring each pod is running 02/25/23 11:42:02.652 +Feb 25 11:42:02.652: INFO: Creating deployment test-cleanup-deployment +STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up 02/25/23 11:42:02.668 [AfterEach] [sig-apps] Deployment test/e2e/apps/deployment.go:84 -Dec 14 17:09:53.672: INFO: Deployment "test-rollover-deployment": -&Deployment{ObjectMeta:{test-rollover-deployment deployment-9963 605b9445-54c6-401b-9e17-8708e45faeb3 31466 2 2022-12-14 17:09:37 +0000 UTC map[name:rollover-pod] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2022-12-14 17:09:39 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:minReadySeconds":{},"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 17:09:51 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003fa3648 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2022-12-14 17:09:37 +0000 UTC,LastTransitionTime:2022-12-14 17:09:37 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rollover-deployment-6d45fd857b" has successfully progressed.,LastUpdateTime:2022-12-14 17:09:51 +0000 UTC,LastTransitionTime:2022-12-14 17:09:37 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} - -Dec 14 17:09:53.676: INFO: New ReplicaSet "test-rollover-deployment-6d45fd857b" of Deployment "test-rollover-deployment": -&ReplicaSet{ObjectMeta:{test-rollover-deployment-6d45fd857b deployment-9963 8d3574b3-2fe1-4af2-8915-9a526fad6da8 31456 2 2022-12-14 17:09:39 +0000 UTC map[name:rollover-pod pod-template-hash:6d45fd857b] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-rollover-deployment 605b9445-54c6-401b-9e17-8708e45faeb3 0xc003fa3c37 0xc003fa3c38}] [] [{kube-controller-manager Update apps/v1 2022-12-14 17:09:39 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"605b9445-54c6-401b-9e17-8708e45faeb3\"}":{}}},"f:spec":{"f:minReadySeconds":{},"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 17:09:51 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 6d45fd857b,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod-template-hash:6d45fd857b] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003fa3ce8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} -Dec 14 17:09:53.676: INFO: All old ReplicaSets of Deployment "test-rollover-deployment": -Dec 14 17:09:53.676: INFO: &ReplicaSet{ObjectMeta:{test-rollover-controller deployment-9963 c2948054-b730-42b0-a58c-c849fe93e9cc 31465 2 2022-12-14 17:09:30 +0000 UTC map[name:rollover-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2] [{apps/v1 Deployment test-rollover-deployment 605b9445-54c6-401b-9e17-8708e45faeb3 0xc003fa39d7 0xc003fa39d8}] [] [{e2e.test Update apps/v1 2022-12-14 17:09:30 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 17:09:51 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"605b9445-54c6-401b-9e17-8708e45faeb3\"}":{}}},"f:spec":{"f:replicas":{}}} } {kube-controller-manager Update apps/v1 2022-12-14 17:09:51 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc003fa3a98 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} -Dec 14 17:09:53.677: INFO: &ReplicaSet{ObjectMeta:{test-rollover-deployment-59b9df946d deployment-9963 f81e86c1-ed2a-40f6-9364-60b0e0933597 31412 2 2022-12-14 17:09:37 +0000 UTC map[name:rollover-pod pod-template-hash:59b9df946d] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-rollover-deployment 605b9445-54c6-401b-9e17-8708e45faeb3 0xc003fa3b07 0xc003fa3b08}] [] [{kube-controller-manager Update apps/v1 2022-12-14 17:09:39 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"605b9445-54c6-401b-9e17-8708e45faeb3\"}":{}}},"f:spec":{"f:minReadySeconds":{},"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"redis-slave\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 17:09:39 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 59b9df946d,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod-template-hash:59b9df946d] map[] [] [] []} {[] [] [{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003fa3bc8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} -Dec 14 17:09:53.681: INFO: Pod "test-rollover-deployment-6d45fd857b-rg7st" is available: -&Pod{ObjectMeta:{test-rollover-deployment-6d45fd857b-rg7st test-rollover-deployment-6d45fd857b- deployment-9963 82702ab1-4e9d-47aa-849d-02b56458aa32 31428 0 2022-12-14 17:09:39 +0000 UTC map[name:rollover-pod pod-template-hash:6d45fd857b] map[] [{apps/v1 ReplicaSet test-rollover-deployment-6d45fd857b 8d3574b3-2fe1-4af2-8915-9a526fad6da8 0xc004088237 0xc004088238}] [] [{kube-controller-manager Update v1 2022-12-14 17:09:39 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8d3574b3-2fe1-4af2-8915-9a526fad6da8\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 17:09:41 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.205\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-427d7,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-427d7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:39 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:39 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.205,StartTime:2022-12-14 17:09:39 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 17:09:40 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,ImageID:registry.k8s.io/e2e-test-images/agnhost@sha256:37dd9ad84b34913b4bc57e00b0a28e8b00f18ca6a5ecec69461aa57402e5c787,ContainerID:cri-o://a1c57a4affd7555320dcc7297664026091e6c4e2b94923187c2000220c1118b1,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.205,},},EphemeralContainerStatuses:[]ContainerStatus{},},} +Feb 25 11:42:02.685: INFO: Deployment "test-cleanup-deployment": +&Deployment{ObjectMeta:{test-cleanup-deployment deployment-2422 5bee2929-834e-4b34-a48e-6ccd1d8ad769 47232 1 2023-02-25 11:42:02 +0000 UTC map[name:cleanup-pod] map[] [] [] [{e2e.test Update apps/v1 2023-02-25 11:42:02 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} }]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:cleanup-pod] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0043e1768 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:0,Replicas:0,UpdatedReplicas:0,AvailableReplicas:0,UnavailableReplicas:0,Conditions:[]DeploymentCondition{},ReadyReplicas:0,CollisionCount:nil,},} + +Feb 25 11:42:02.692: INFO: New ReplicaSet of Deployment "test-cleanup-deployment" is nil. [AfterEach] [sig-apps] Deployment test/e2e/framework/framework.go:187 -Dec 14 17:09:53.681: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "deployment-9963" for this suite. 12/14/22 17:09:53.688 -{"msg":"PASSED [sig-apps] Deployment deployment should support rollover [Conformance]","completed":351,"skipped":6542,"failed":0} +Feb 25 11:42:02.700: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "deployment-2422" for this suite. 02/25/23 11:42:02.708 +{"msg":"PASSED [sig-apps] Deployment deployment should delete old replica sets [Conformance]","completed":349,"skipped":6507,"failed":0} ------------------------------ -• [SLOW TEST] [23.250 seconds] +• [SLOW TEST] [5.141 seconds] [sig-apps] Deployment test/e2e/apps/framework.go:23 - deployment should support rollover [Conformance] - test/e2e/apps/deployment.go:132 + deployment should delete old replica sets [Conformance] + test/e2e/apps/deployment.go:122 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-apps] Deployment test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:09:30.451 - Dec 14 17:09:30.451: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename deployment 12/14/22 17:09:30.452 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:30.491 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:30.494 + STEP: Creating a kubernetes client 02/25/23 11:41:57.578 + Feb 25 11:41:57.579: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename deployment 02/25/23 11:41:57.581 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:41:57.607 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:41:57.611 [BeforeEach] [sig-apps] Deployment test/e2e/apps/deployment.go:91 - [It] deployment should support rollover [Conformance] - test/e2e/apps/deployment.go:132 - Dec 14 17:09:30.527: INFO: Pod name rollover-pod: Found 0 pods out of 1 - Dec 14 17:09:35.532: INFO: Pod name rollover-pod: Found 1 pods out of 1 - STEP: ensuring each pod is running 12/14/22 17:09:35.532 - Dec 14 17:09:35.533: INFO: Waiting for pods owned by replica set "test-rollover-controller" to become ready - Dec 14 17:09:37.543: INFO: Creating deployment "test-rollover-deployment" - Dec 14 17:09:37.560: INFO: Make sure deployment "test-rollover-deployment" performs scaling operations - Dec 14 17:09:39.574: INFO: Check revision of new replica set for deployment "test-rollover-deployment" - Dec 14 17:09:39.584: INFO: Ensure that both replica sets have 1 created replica - Dec 14 17:09:39.593: INFO: Rollover old replica sets for deployment "test-rollover-deployment" with new image update - Dec 14 17:09:39.610: INFO: Updating deployment test-rollover-deployment - Dec 14 17:09:39.610: INFO: Wait deployment "test-rollover-deployment" to be observed by the deployment controller - Dec 14 17:09:41.622: INFO: Wait for revision update of deployment "test-rollover-deployment" to 2 - Dec 14 17:09:41.634: INFO: Make sure deployment "test-rollover-deployment" is complete - Dec 14 17:09:41.644: INFO: all replica sets need to contain the pod-template-hash label - Dec 14 17:09:41.644: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 41, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 17:09:43.658: INFO: all replica sets need to contain the pod-template-hash label - Dec 14 17:09:43.658: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 41, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 17:09:45.656: INFO: all replica sets need to contain the pod-template-hash label - Dec 14 17:09:45.656: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 41, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 17:09:47.658: INFO: all replica sets need to contain the pod-template-hash label - Dec 14 17:09:47.658: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 41, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 17:09:49.658: INFO: all replica sets need to contain the pod-template-hash label - Dec 14 17:09:49.659: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:1, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 41, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 17:09:51.685: INFO: - Dec 14 17:09:51.685: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:2, Replicas:2, UpdatedReplicas:1, ReadyReplicas:2, AvailableReplicas:2, UnavailableReplicas:0, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"MinimumReplicasAvailable", Message:"Deployment has minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 51, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 37, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"test-rollover-deployment-6d45fd857b\" is progressing."}}, CollisionCount:(*int32)(nil)} - Dec 14 17:09:53.659: INFO: - Dec 14 17:09:53.659: INFO: Ensure that both old replica sets have no replicas + [It] deployment should delete old replica sets [Conformance] + test/e2e/apps/deployment.go:122 + Feb 25 11:41:57.645: INFO: Pod name cleanup-pod: Found 0 pods out of 1 + Feb 25 11:42:02.652: INFO: Pod name cleanup-pod: Found 1 pods out of 1 + STEP: ensuring each pod is running 02/25/23 11:42:02.652 + Feb 25 11:42:02.652: INFO: Creating deployment test-cleanup-deployment + STEP: Waiting for deployment test-cleanup-deployment history to be cleaned up 02/25/23 11:42:02.668 [AfterEach] [sig-apps] Deployment test/e2e/apps/deployment.go:84 - Dec 14 17:09:53.672: INFO: Deployment "test-rollover-deployment": - &Deployment{ObjectMeta:{test-rollover-deployment deployment-9963 605b9445-54c6-401b-9e17-8708e45faeb3 31466 2 2022-12-14 17:09:37 +0000 UTC map[name:rollover-pod] map[deployment.kubernetes.io/revision:2] [] [] [{e2e.test Update apps/v1 2022-12-14 17:09:39 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:minReadySeconds":{},"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 17:09:51 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/revision":{}}},"f:status":{"f:availableReplicas":{},"f:conditions":{".":{},"k:{\"type\":\"Available\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Progressing\"}":{".":{},"f:lastTransitionTime":{},"f:lastUpdateTime":{},"f:message":{},"f:reason":{},"f:status":{},"f:type":{}}},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{},"f:updatedReplicas":{}}} status}]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003fa3648 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:0,MaxSurge:1,},},MinReadySeconds:10,RevisionHistoryLimit:*10,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:2,Replicas:1,UpdatedReplicas:1,AvailableReplicas:1,UnavailableReplicas:0,Conditions:[]DeploymentCondition{DeploymentCondition{Type:Available,Status:True,Reason:MinimumReplicasAvailable,Message:Deployment has minimum availability.,LastUpdateTime:2022-12-14 17:09:37 +0000 UTC,LastTransitionTime:2022-12-14 17:09:37 +0000 UTC,},DeploymentCondition{Type:Progressing,Status:True,Reason:NewReplicaSetAvailable,Message:ReplicaSet "test-rollover-deployment-6d45fd857b" has successfully progressed.,LastUpdateTime:2022-12-14 17:09:51 +0000 UTC,LastTransitionTime:2022-12-14 17:09:37 +0000 UTC,},},ReadyReplicas:1,CollisionCount:nil,},} - - Dec 14 17:09:53.676: INFO: New ReplicaSet "test-rollover-deployment-6d45fd857b" of Deployment "test-rollover-deployment": - &ReplicaSet{ObjectMeta:{test-rollover-deployment-6d45fd857b deployment-9963 8d3574b3-2fe1-4af2-8915-9a526fad6da8 31456 2 2022-12-14 17:09:39 +0000 UTC map[name:rollover-pod pod-template-hash:6d45fd857b] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:2] [{apps/v1 Deployment test-rollover-deployment 605b9445-54c6-401b-9e17-8708e45faeb3 0xc003fa3c37 0xc003fa3c38}] [] [{kube-controller-manager Update apps/v1 2022-12-14 17:09:39 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"605b9445-54c6-401b-9e17-8708e45faeb3\"}":{}}},"f:spec":{"f:minReadySeconds":{},"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 17:09:51 +0000 UTC FieldsV1 {"f:status":{"f:availableReplicas":{},"f:fullyLabeledReplicas":{},"f:observedGeneration":{},"f:readyReplicas":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 6d45fd857b,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod-template-hash:6d45fd857b] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003fa3ce8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:1,FullyLabeledReplicas:1,ObservedGeneration:2,ReadyReplicas:1,AvailableReplicas:1,Conditions:[]ReplicaSetCondition{},},} - Dec 14 17:09:53.676: INFO: All old ReplicaSets of Deployment "test-rollover-deployment": - Dec 14 17:09:53.676: INFO: &ReplicaSet{ObjectMeta:{test-rollover-controller deployment-9963 c2948054-b730-42b0-a58c-c849fe93e9cc 31465 2 2022-12-14 17:09:30 +0000 UTC map[name:rollover-pod pod:httpd] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2] [{apps/v1 Deployment test-rollover-deployment 605b9445-54c6-401b-9e17-8708e45faeb3 0xc003fa39d7 0xc003fa39d8}] [] [{e2e.test Update apps/v1 2022-12-14 17:09:30 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"httpd\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 17:09:51 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"605b9445-54c6-401b-9e17-8708e45faeb3\"}":{}}},"f:spec":{"f:replicas":{}}} } {kube-controller-manager Update apps/v1 2022-12-14 17:09:51 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod: httpd,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod:httpd] map[] [] [] []} {[] [] [{httpd registry.k8s.io/e2e-test-images/httpd:2.4.38-2 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent nil false false false}] [] Always 0xc003fa3a98 ClusterFirst map[] false false false PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:0,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} - Dec 14 17:09:53.677: INFO: &ReplicaSet{ObjectMeta:{test-rollover-deployment-59b9df946d deployment-9963 f81e86c1-ed2a-40f6-9364-60b0e0933597 31412 2 2022-12-14 17:09:37 +0000 UTC map[name:rollover-pod pod-template-hash:59b9df946d] map[deployment.kubernetes.io/desired-replicas:1 deployment.kubernetes.io/max-replicas:2 deployment.kubernetes.io/revision:1] [{apps/v1 Deployment test-rollover-deployment 605b9445-54c6-401b-9e17-8708e45faeb3 0xc003fa3b07 0xc003fa3b08}] [] [{kube-controller-manager Update apps/v1 2022-12-14 17:09:39 +0000 UTC FieldsV1 {"f:metadata":{"f:annotations":{".":{},"f:deployment.kubernetes.io/desired-replicas":{},"f:deployment.kubernetes.io/max-replicas":{},"f:deployment.kubernetes.io/revision":{}},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"605b9445-54c6-401b-9e17-8708e45faeb3\"}":{}}},"f:spec":{"f:minReadySeconds":{},"f:replicas":{},"f:selector":{},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"redis-slave\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} } {kube-controller-manager Update apps/v1 2022-12-14 17:09:39 +0000 UTC FieldsV1 {"f:status":{"f:observedGeneration":{},"f:replicas":{}}} status}]},Spec:ReplicaSetSpec{Replicas:*0,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: rollover-pod,pod-template-hash: 59b9df946d,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:rollover-pod pod-template-hash:59b9df946d] map[] [] [] []} {[] [] [{redis-slave gcr.io/google_samples/gb-redisslave:nonexistent [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc003fa3bc8 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},MinReadySeconds:10,},Status:ReplicaSetStatus{Replicas:0,FullyLabeledReplicas:0,ObservedGeneration:2,ReadyReplicas:0,AvailableReplicas:0,Conditions:[]ReplicaSetCondition{},},} - Dec 14 17:09:53.681: INFO: Pod "test-rollover-deployment-6d45fd857b-rg7st" is available: - &Pod{ObjectMeta:{test-rollover-deployment-6d45fd857b-rg7st test-rollover-deployment-6d45fd857b- deployment-9963 82702ab1-4e9d-47aa-849d-02b56458aa32 31428 0 2022-12-14 17:09:39 +0000 UTC map[name:rollover-pod pod-template-hash:6d45fd857b] map[] [{apps/v1 ReplicaSet test-rollover-deployment-6d45fd857b 8d3574b3-2fe1-4af2-8915-9a526fad6da8 0xc004088237 0xc004088238}] [] [{kube-controller-manager Update v1 2022-12-14 17:09:39 +0000 UTC FieldsV1 {"f:metadata":{"f:generateName":{},"f:labels":{".":{},"f:name":{},"f:pod-template-hash":{}},"f:ownerReferences":{".":{},"k:{\"uid\":\"8d3574b3-2fe1-4af2-8915-9a526fad6da8\"}":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:enableServiceLinks":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}} } {kubelet Update v1 2022-12-14 17:09:41 +0000 UTC FieldsV1 {"f:status":{"f:conditions":{"k:{\"type\":\"ContainersReady\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Initialized\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}},"k:{\"type\":\"Ready\"}":{".":{},"f:lastProbeTime":{},"f:lastTransitionTime":{},"f:status":{},"f:type":{}}},"f:containerStatuses":{},"f:hostIP":{},"f:phase":{},"f:podIP":{},"f:podIPs":{".":{},"k:{\"ip\":\"10.233.66.205\"}":{".":{},"f:ip":{}}},"f:startTime":{}}} status}]},Spec:PodSpec{Volumes:[]Volume{Volume{Name:kube-api-access-427d7,VolumeSource:VolumeSource{HostPath:nil,EmptyDir:nil,GCEPersistentDisk:nil,AWSElasticBlockStore:nil,GitRepo:nil,Secret:nil,NFS:nil,ISCSI:nil,Glusterfs:nil,PersistentVolumeClaim:nil,RBD:nil,FlexVolume:nil,Cinder:nil,CephFS:nil,Flocker:nil,DownwardAPI:nil,FC:nil,AzureFile:nil,ConfigMap:nil,VsphereVolume:nil,Quobyte:nil,AzureDisk:nil,PhotonPersistentDisk:nil,PortworxVolume:nil,ScaleIO:nil,Projected:&ProjectedVolumeSource{Sources:[]VolumeProjection{VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:nil,ServiceAccountToken:&ServiceAccountTokenProjection{Audience:,ExpirationSeconds:*3607,Path:token,},},VolumeProjection{Secret:nil,DownwardAPI:nil,ConfigMap:&ConfigMapProjection{LocalObjectReference:LocalObjectReference{Name:kube-root-ca.crt,},Items:[]KeyToPath{KeyToPath{Key:ca.crt,Path:ca.crt,Mode:nil,},},Optional:nil,},ServiceAccountToken:nil,},VolumeProjection{Secret:nil,DownwardAPI:&DownwardAPIProjection{Items:[]DownwardAPIVolumeFile{DownwardAPIVolumeFile{Path:namespace,FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,Mode:nil,},},},ConfigMap:nil,ServiceAccountToken:nil,},},DefaultMode:*420,},StorageOS:nil,CSI:nil,Ephemeral:nil,},},},Containers:[]Container{Container{Name:agnhost,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-427d7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,},},RestartPolicy:Always,TerminationGracePeriodSeconds:*0,ActiveDeadlineSeconds:nil,DNSPolicy:ClusterFirst,NodeSelector:map[string]string{},ServiceAccountName:default,DeprecatedServiceAccount:default,NodeName:kiedai4iakae-3,HostNetwork:false,HostPID:false,HostIPC:false,SecurityContext:&PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,},ImagePullSecrets:[]LocalObjectReference{},Hostname:,Subdomain:,Affinity:nil,SchedulerName:default-scheduler,InitContainers:[]Container{},AutomountServiceAccountToken:nil,Tolerations:[]Toleration{Toleration{Key:node.kubernetes.io/not-ready,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},Toleration{Key:node.kubernetes.io/unreachable,Operator:Exists,Value:,Effect:NoExecute,TolerationSeconds:*300,},},HostAliases:[]HostAlias{},PriorityClassName:,Priority:*0,DNSConfig:nil,ShareProcessNamespace:nil,ReadinessGates:[]PodReadinessGate{},RuntimeClassName:nil,EnableServiceLinks:*true,PreemptionPolicy:*PreemptLowerPriority,Overhead:ResourceList{},TopologySpreadConstraints:[]TopologySpreadConstraint{},EphemeralContainers:[]EphemeralContainer{},SetHostnameAsFQDN:nil,OS:nil,HostUsers:nil,},Status:PodStatus{Phase:Running,Conditions:[]PodCondition{PodCondition{Type:Initialized,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:39 +0000 UTC,Reason:,Message:,},PodCondition{Type:Ready,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:ContainersReady,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:41 +0000 UTC,Reason:,Message:,},PodCondition{Type:PodScheduled,Status:True,LastProbeTime:0001-01-01 00:00:00 +0000 UTC,LastTransitionTime:2022-12-14 17:09:39 +0000 UTC,Reason:,Message:,},},Message:,Reason:,HostIP:192.168.121.244,PodIP:10.233.66.205,StartTime:2022-12-14 17:09:39 +0000 UTC,ContainerStatuses:[]ContainerStatus{ContainerStatus{Name:agnhost,State:ContainerState{Waiting:nil,Running:&ContainerStateRunning{StartedAt:2022-12-14 17:09:40 +0000 UTC,},Terminated:nil,},LastTerminationState:ContainerState{Waiting:nil,Running:nil,Terminated:nil,},Ready:true,RestartCount:0,Image:registry.k8s.io/e2e-test-images/agnhost:2.40,ImageID:registry.k8s.io/e2e-test-images/agnhost@sha256:37dd9ad84b34913b4bc57e00b0a28e8b00f18ca6a5ecec69461aa57402e5c787,ContainerID:cri-o://a1c57a4affd7555320dcc7297664026091e6c4e2b94923187c2000220c1118b1,Started:*true,},},QOSClass:BestEffort,InitContainerStatuses:[]ContainerStatus{},NominatedNodeName:,PodIPs:[]PodIP{PodIP{IP:10.233.66.205,},},EphemeralContainerStatuses:[]ContainerStatus{},},} + Feb 25 11:42:02.685: INFO: Deployment "test-cleanup-deployment": + &Deployment{ObjectMeta:{test-cleanup-deployment deployment-2422 5bee2929-834e-4b34-a48e-6ccd1d8ad769 47232 1 2023-02-25 11:42:02 +0000 UTC map[name:cleanup-pod] map[] [] [] [{e2e.test Update apps/v1 2023-02-25 11:42:02 +0000 UTC FieldsV1 {"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:progressDeadlineSeconds":{},"f:replicas":{},"f:revisionHistoryLimit":{},"f:selector":{},"f:strategy":{"f:rollingUpdate":{".":{},"f:maxSurge":{},"f:maxUnavailable":{}},"f:type":{}},"f:template":{"f:metadata":{"f:labels":{".":{},"f:name":{}}},"f:spec":{"f:containers":{"k:{\"name\":\"agnhost\"}":{".":{},"f:image":{},"f:imagePullPolicy":{},"f:name":{},"f:resources":{},"f:securityContext":{},"f:terminationMessagePath":{},"f:terminationMessagePolicy":{}}},"f:dnsPolicy":{},"f:restartPolicy":{},"f:schedulerName":{},"f:securityContext":{},"f:terminationGracePeriodSeconds":{}}}}} }]},Spec:DeploymentSpec{Replicas:*1,Selector:&v1.LabelSelector{MatchLabels:map[string]string{name: cleanup-pod,},MatchExpressions:[]LabelSelectorRequirement{},},Template:{{ 0 0001-01-01 00:00:00 +0000 UTC map[name:cleanup-pod] map[] [] [] []} {[] [] [{agnhost registry.k8s.io/e2e-test-images/agnhost:2.40 [] [] [] [] [] {map[] map[]} [] [] nil nil nil nil /dev/termination-log File IfNotPresent SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,} false false false}] [] Always 0xc0043e1768 ClusterFirst map[] false false false &PodSecurityContext{SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,SupplementalGroups:[],FSGroup:nil,RunAsGroup:nil,Sysctls:[]Sysctl{},WindowsOptions:nil,FSGroupChangePolicy:nil,SeccompProfile:nil,} [] nil default-scheduler [] [] nil [] map[] [] nil }},Strategy:DeploymentStrategy{Type:RollingUpdate,RollingUpdate:&RollingUpdateDeployment{MaxUnavailable:25%!,(MISSING)MaxSurge:25%!,(MISSING)},},MinReadySeconds:0,RevisionHistoryLimit:*0,Paused:false,ProgressDeadlineSeconds:*600,},Status:DeploymentStatus{ObservedGeneration:0,Replicas:0,UpdatedReplicas:0,AvailableReplicas:0,UnavailableReplicas:0,Conditions:[]DeploymentCondition{},ReadyReplicas:0,CollisionCount:nil,},} + + Feb 25 11:42:02.692: INFO: New ReplicaSet of Deployment "test-cleanup-deployment" is nil. [AfterEach] [sig-apps] Deployment test/e2e/framework/framework.go:187 - Dec 14 17:09:53.681: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "deployment-9963" for this suite. 12/14/22 17:09:53.688 + Feb 25 11:42:02.700: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "deployment-2422" for this suite. 02/25/23 11:42:02.708 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSS +SSSSSSSSSS ------------------------------ -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - should mutate pod and apply defaults after mutation [Conformance] - test/e2e/apimachinery/webhook.go:263 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +[sig-node] RuntimeClass + should support RuntimeClasses API operations [Conformance] + test/e2e/common/node/runtimeclass.go:189 +[BeforeEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:09:53.706 -Dec 14 17:09:53.706: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename webhook 12/14/22 17:09:53.712 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:53.736 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:53.739 -[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 -STEP: Setting up server cert 12/14/22 17:09:53.793 -STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 17:09:54.986 -STEP: Deploying the webhook pod 12/14/22 17:09:55.004 -STEP: Wait for the deployment to be ready 12/14/22 17:09:55.021 -Dec 14 17:09:55.033: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set -Dec 14 17:09:57.050: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 55, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 55, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 55, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 55, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} -STEP: Deploying the webhook service 12/14/22 17:09:59.055 -STEP: Verifying the service has paired with the endpoint 12/14/22 17:09:59.074 -Dec 14 17:10:00.074: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 -[It] should mutate pod and apply defaults after mutation [Conformance] - test/e2e/apimachinery/webhook.go:263 -STEP: Registering the mutating pod webhook via the AdmissionRegistration API 12/14/22 17:10:00.082 -STEP: create a pod that should be updated by the webhook 12/14/22 17:10:00.128 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +STEP: Creating a kubernetes client 02/25/23 11:42:02.724 +Feb 25 11:42:02.725: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename runtimeclass 02/25/23 11:42:02.727 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:02.778 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:02.786 +[It] should support RuntimeClasses API operations [Conformance] + test/e2e/common/node/runtimeclass.go:189 +STEP: getting /apis 02/25/23 11:42:02.79 +STEP: getting /apis/node.k8s.io 02/25/23 11:42:02.794 +STEP: getting /apis/node.k8s.io/v1 02/25/23 11:42:02.796 +STEP: creating 02/25/23 11:42:02.798 +STEP: watching 02/25/23 11:42:02.828 +Feb 25 11:42:02.829: INFO: starting watch +STEP: getting 02/25/23 11:42:02.838 +STEP: listing 02/25/23 11:42:02.845 +STEP: patching 02/25/23 11:42:02.855 +STEP: updating 02/25/23 11:42:02.863 +Feb 25 11:42:02.872: INFO: waiting for watch events with expected annotations +STEP: deleting 02/25/23 11:42:02.872 +STEP: deleting a collection 02/25/23 11:42:02.929 +[AfterEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:187 -Dec 14 17:10:00.182: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "webhook-6504" for this suite. 12/14/22 17:10:00.199 -STEP: Destroying namespace "webhook-6504-markers" for this suite. 12/14/22 17:10:00.213 -[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 -{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should mutate pod and apply defaults after mutation [Conformance]","completed":352,"skipped":6554,"failed":0} +Feb 25 11:42:02.984: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "runtimeclass-2784" for this suite. 02/25/23 11:42:02.992 +{"msg":"PASSED [sig-node] RuntimeClass should support RuntimeClasses API operations [Conformance]","completed":350,"skipped":6517,"failed":0} ------------------------------ -• [SLOW TEST] [6.645 seconds] -[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] -test/e2e/apimachinery/framework.go:23 - should mutate pod and apply defaults after mutation [Conformance] - test/e2e/apimachinery/webhook.go:263 +• [0.286 seconds] +[sig-node] RuntimeClass +test/e2e/common/node/framework.go:23 + should support RuntimeClasses API operations [Conformance] + test/e2e/common/node/runtimeclass.go:189 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + [BeforeEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:09:53.706 - Dec 14 17:09:53.706: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename webhook 12/14/22 17:09:53.712 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:09:53.736 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:09:53.739 - [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:89 - STEP: Setting up server cert 12/14/22 17:09:53.793 - STEP: Create role binding to let webhook read extension-apiserver-authentication 12/14/22 17:09:54.986 - STEP: Deploying the webhook pod 12/14/22 17:09:55.004 - STEP: Wait for the deployment to be ready 12/14/22 17:09:55.021 - Dec 14 17:09:55.033: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set - Dec 14 17:09:57.050: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 55, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 55, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2022, time.December, 14, 17, 9, 55, 0, time.Local), LastTransitionTime:time.Date(2022, time.December, 14, 17, 9, 55, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} - STEP: Deploying the webhook service 12/14/22 17:09:59.055 - STEP: Verifying the service has paired with the endpoint 12/14/22 17:09:59.074 - Dec 14 17:10:00.074: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 - [It] should mutate pod and apply defaults after mutation [Conformance] - test/e2e/apimachinery/webhook.go:263 - STEP: Registering the mutating pod webhook via the AdmissionRegistration API 12/14/22 17:10:00.082 - STEP: create a pod that should be updated by the webhook 12/14/22 17:10:00.128 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + STEP: Creating a kubernetes client 02/25/23 11:42:02.724 + Feb 25 11:42:02.725: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename runtimeclass 02/25/23 11:42:02.727 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:02.778 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:02.786 + [It] should support RuntimeClasses API operations [Conformance] + test/e2e/common/node/runtimeclass.go:189 + STEP: getting /apis 02/25/23 11:42:02.79 + STEP: getting /apis/node.k8s.io 02/25/23 11:42:02.794 + STEP: getting /apis/node.k8s.io/v1 02/25/23 11:42:02.796 + STEP: creating 02/25/23 11:42:02.798 + STEP: watching 02/25/23 11:42:02.828 + Feb 25 11:42:02.829: INFO: starting watch + STEP: getting 02/25/23 11:42:02.838 + STEP: listing 02/25/23 11:42:02.845 + STEP: patching 02/25/23 11:42:02.855 + STEP: updating 02/25/23 11:42:02.863 + Feb 25 11:42:02.872: INFO: waiting for watch events with expected annotations + STEP: deleting 02/25/23 11:42:02.872 + STEP: deleting a collection 02/25/23 11:42:02.929 + [AfterEach] [sig-node] RuntimeClass test/e2e/framework/framework.go:187 - Dec 14 17:10:00.182: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "webhook-6504" for this suite. 12/14/22 17:10:00.199 - STEP: Destroying namespace "webhook-6504-markers" for this suite. 12/14/22 17:10:00.213 - [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] - test/e2e/apimachinery/webhook.go:104 + Feb 25 11:42:02.984: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "runtimeclass-2784" for this suite. 02/25/23 11:42:02.992 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSS ------------------------------- -[sig-cli] Kubectl client Update Demo - should scale a replication controller [Conformance] - test/e2e/kubectl/kubectl.go:350 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:10:00.353 -Dec 14 17:10:00.354: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename kubectl 12/14/22 17:10:00.359 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:10:00.391 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:10:00.394 -[BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 -[BeforeEach] Update Demo - test/e2e/kubectl/kubectl.go:324 -[It] should scale a replication controller [Conformance] - test/e2e/kubectl/kubectl.go:350 -STEP: creating a replication controller 12/14/22 17:10:00.397 -Dec 14 17:10:00.398: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 create -f -' -Dec 14 17:10:00.943: INFO: stderr: "" -Dec 14 17:10:00.943: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" -STEP: waiting for all containers in name=update-demo pods to come up. 12/14/22 17:10:00.943 -Dec 14 17:10:00.944: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' -Dec 14 17:10:01.099: INFO: stderr: "" -Dec 14 17:10:01.099: INFO: stdout: "update-demo-nautilus-sp2nw update-demo-nautilus-twj2k " -Dec 14 17:10:01.100: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-sp2nw -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' -Dec 14 17:10:01.232: INFO: stderr: "" -Dec 14 17:10:01.232: INFO: stdout: "" -Dec 14 17:10:01.232: INFO: update-demo-nautilus-sp2nw is created but not running -Dec 14 17:10:06.234: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' -Dec 14 17:10:06.399: INFO: stderr: "" -Dec 14 17:10:06.399: INFO: stdout: "update-demo-nautilus-sp2nw update-demo-nautilus-twj2k " -Dec 14 17:10:06.400: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-sp2nw -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' -Dec 14 17:10:06.559: INFO: stderr: "" -Dec 14 17:10:06.559: INFO: stdout: "true" -Dec 14 17:10:06.559: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-sp2nw -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' -Dec 14 17:10:06.713: INFO: stderr: "" -Dec 14 17:10:06.714: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" -Dec 14 17:10:06.714: INFO: validating pod update-demo-nautilus-sp2nw -Dec 14 17:10:06.725: INFO: got data: { - "image": "nautilus.jpg" -} - -Dec 14 17:10:06.725: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -Dec 14 17:10:06.725: INFO: update-demo-nautilus-sp2nw is verified up and running -Dec 14 17:10:06.726: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-twj2k -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' -Dec 14 17:10:06.859: INFO: stderr: "" -Dec 14 17:10:06.859: INFO: stdout: "true" -Dec 14 17:10:06.859: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-twj2k -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' -Dec 14 17:10:07.022: INFO: stderr: "" -Dec 14 17:10:07.022: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" -Dec 14 17:10:07.022: INFO: validating pod update-demo-nautilus-twj2k -Dec 14 17:10:22.401: INFO: got data: { - "image": "nautilus.jpg" -} - -Dec 14 17:10:22.401: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -Dec 14 17:10:22.401: INFO: update-demo-nautilus-twj2k is verified up and running -STEP: scaling down the replication controller 12/14/22 17:10:22.401 -Dec 14 17:10:22.425: INFO: scanned /root for discovery docs: -Dec 14 17:10:22.426: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 scale rc update-demo-nautilus --replicas=1 --timeout=5m' -Dec 14 17:10:23.615: INFO: stderr: "" -Dec 14 17:10:23.615: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" -STEP: waiting for all containers in name=update-demo pods to come up. 12/14/22 17:10:23.615 -Dec 14 17:10:23.615: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' -Dec 14 17:10:23.735: INFO: stderr: "" -Dec 14 17:10:23.736: INFO: stdout: "update-demo-nautilus-sp2nw " -Dec 14 17:10:23.736: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-sp2nw -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' -Dec 14 17:10:23.890: INFO: stderr: "" -Dec 14 17:10:23.890: INFO: stdout: "true" -Dec 14 17:10:23.890: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-sp2nw -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' -Dec 14 17:10:24.033: INFO: stderr: "" -Dec 14 17:10:24.033: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" -Dec 14 17:10:24.033: INFO: validating pod update-demo-nautilus-sp2nw -Dec 14 17:10:24.045: INFO: got data: { - "image": "nautilus.jpg" -} - -Dec 14 17:10:24.045: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -Dec 14 17:10:24.045: INFO: update-demo-nautilus-sp2nw is verified up and running -STEP: scaling up the replication controller 12/14/22 17:10:24.045 -Dec 14 17:10:24.063: INFO: scanned /root for discovery docs: -Dec 14 17:10:24.063: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 scale rc update-demo-nautilus --replicas=2 --timeout=5m' -Dec 14 17:10:25.280: INFO: stderr: "" -Dec 14 17:10:25.280: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" -STEP: waiting for all containers in name=update-demo pods to come up. 12/14/22 17:10:25.28 -Dec 14 17:10:25.280: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' -Dec 14 17:10:25.456: INFO: stderr: "" -Dec 14 17:10:25.456: INFO: stdout: "update-demo-nautilus-nfpw2 update-demo-nautilus-sp2nw " -Dec 14 17:10:25.456: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-nfpw2 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' -Dec 14 17:10:25.658: INFO: stderr: "" -Dec 14 17:10:25.658: INFO: stdout: "" -Dec 14 17:10:25.659: INFO: update-demo-nautilus-nfpw2 is created but not running -Dec 14 17:10:30.659: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' -Dec 14 17:10:30.806: INFO: stderr: "" -Dec 14 17:10:30.806: INFO: stdout: "update-demo-nautilus-nfpw2 update-demo-nautilus-sp2nw " -Dec 14 17:10:30.807: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-nfpw2 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' -Dec 14 17:10:31.011: INFO: stderr: "" -Dec 14 17:10:31.011: INFO: stdout: "true" -Dec 14 17:10:31.011: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-nfpw2 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' -Dec 14 17:10:31.167: INFO: stderr: "" -Dec 14 17:10:31.167: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" -Dec 14 17:10:31.167: INFO: validating pod update-demo-nautilus-nfpw2 -Dec 14 17:10:31.182: INFO: got data: { - "image": "nautilus.jpg" -} - -Dec 14 17:10:31.182: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -Dec 14 17:10:31.182: INFO: update-demo-nautilus-nfpw2 is verified up and running -Dec 14 17:10:31.182: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-sp2nw -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' -Dec 14 17:10:31.339: INFO: stderr: "" -Dec 14 17:10:31.339: INFO: stdout: "true" -Dec 14 17:10:31.340: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-sp2nw -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' -Dec 14 17:10:31.454: INFO: stderr: "" -Dec 14 17:10:31.454: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" -Dec 14 17:10:31.454: INFO: validating pod update-demo-nautilus-sp2nw -Dec 14 17:10:31.464: INFO: got data: { - "image": "nautilus.jpg" -} - -Dec 14 17:10:31.464: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . -Dec 14 17:10:31.464: INFO: update-demo-nautilus-sp2nw is verified up and running -STEP: using delete to clean up resources 12/14/22 17:10:31.464 -Dec 14 17:10:31.464: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 delete --grace-period=0 --force -f -' -Dec 14 17:10:31.602: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" -Dec 14 17:10:31.602: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" -Dec 14 17:10:31.603: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get rc,svc -l name=update-demo --no-headers' -Dec 14 17:10:31.745: INFO: stderr: "No resources found in kubectl-3777 namespace.\n" -Dec 14 17:10:31.745: INFO: stdout: "" -Dec 14 17:10:31.745: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods -l name=update-demo -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' -Dec 14 17:10:32.009: INFO: stderr: "" -Dec 14 17:10:32.009: INFO: stdout: "" -[AfterEach] [sig-cli] Kubectl client - test/e2e/framework/framework.go:187 -Dec 14 17:10:32.010: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "kubectl-3777" for this suite. 12/14/22 17:10:32.017 -{"msg":"PASSED [sig-cli] Kubectl client Update Demo should scale a replication controller [Conformance]","completed":353,"skipped":6562,"failed":0} ------------------------------- -• [SLOW TEST] [31.673 seconds] -[sig-cli] Kubectl client -test/e2e/kubectl/framework.go:23 - Update Demo - test/e2e/kubectl/kubectl.go:322 - should scale a replication controller [Conformance] - test/e2e/kubectl/kubectl.go:350 - - Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-cli] Kubectl client - test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:10:00.353 - Dec 14 17:10:00.354: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename kubectl 12/14/22 17:10:00.359 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:10:00.391 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:10:00.394 - [BeforeEach] [sig-cli] Kubectl client - test/e2e/kubectl/kubectl.go:272 - [BeforeEach] Update Demo - test/e2e/kubectl/kubectl.go:324 - [It] should scale a replication controller [Conformance] - test/e2e/kubectl/kubectl.go:350 - STEP: creating a replication controller 12/14/22 17:10:00.397 - Dec 14 17:10:00.398: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 create -f -' - Dec 14 17:10:00.943: INFO: stderr: "" - Dec 14 17:10:00.943: INFO: stdout: "replicationcontroller/update-demo-nautilus created\n" - STEP: waiting for all containers in name=update-demo pods to come up. 12/14/22 17:10:00.943 - Dec 14 17:10:00.944: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' - Dec 14 17:10:01.099: INFO: stderr: "" - Dec 14 17:10:01.099: INFO: stdout: "update-demo-nautilus-sp2nw update-demo-nautilus-twj2k " - Dec 14 17:10:01.100: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-sp2nw -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' - Dec 14 17:10:01.232: INFO: stderr: "" - Dec 14 17:10:01.232: INFO: stdout: "" - Dec 14 17:10:01.232: INFO: update-demo-nautilus-sp2nw is created but not running - Dec 14 17:10:06.234: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' - Dec 14 17:10:06.399: INFO: stderr: "" - Dec 14 17:10:06.399: INFO: stdout: "update-demo-nautilus-sp2nw update-demo-nautilus-twj2k " - Dec 14 17:10:06.400: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-sp2nw -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' - Dec 14 17:10:06.559: INFO: stderr: "" - Dec 14 17:10:06.559: INFO: stdout: "true" - Dec 14 17:10:06.559: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-sp2nw -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' - Dec 14 17:10:06.713: INFO: stderr: "" - Dec 14 17:10:06.714: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" - Dec 14 17:10:06.714: INFO: validating pod update-demo-nautilus-sp2nw - Dec 14 17:10:06.725: INFO: got data: { - "image": "nautilus.jpg" - } - - Dec 14 17:10:06.725: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . - Dec 14 17:10:06.725: INFO: update-demo-nautilus-sp2nw is verified up and running - Dec 14 17:10:06.726: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-twj2k -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' - Dec 14 17:10:06.859: INFO: stderr: "" - Dec 14 17:10:06.859: INFO: stdout: "true" - Dec 14 17:10:06.859: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-twj2k -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' - Dec 14 17:10:07.022: INFO: stderr: "" - Dec 14 17:10:07.022: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" - Dec 14 17:10:07.022: INFO: validating pod update-demo-nautilus-twj2k - Dec 14 17:10:22.401: INFO: got data: { - "image": "nautilus.jpg" - } - - Dec 14 17:10:22.401: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . - Dec 14 17:10:22.401: INFO: update-demo-nautilus-twj2k is verified up and running - STEP: scaling down the replication controller 12/14/22 17:10:22.401 - Dec 14 17:10:22.425: INFO: scanned /root for discovery docs: - Dec 14 17:10:22.426: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 scale rc update-demo-nautilus --replicas=1 --timeout=5m' - Dec 14 17:10:23.615: INFO: stderr: "" - Dec 14 17:10:23.615: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" - STEP: waiting for all containers in name=update-demo pods to come up. 12/14/22 17:10:23.615 - Dec 14 17:10:23.615: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' - Dec 14 17:10:23.735: INFO: stderr: "" - Dec 14 17:10:23.736: INFO: stdout: "update-demo-nautilus-sp2nw " - Dec 14 17:10:23.736: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-sp2nw -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' - Dec 14 17:10:23.890: INFO: stderr: "" - Dec 14 17:10:23.890: INFO: stdout: "true" - Dec 14 17:10:23.890: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-sp2nw -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' - Dec 14 17:10:24.033: INFO: stderr: "" - Dec 14 17:10:24.033: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" - Dec 14 17:10:24.033: INFO: validating pod update-demo-nautilus-sp2nw - Dec 14 17:10:24.045: INFO: got data: { - "image": "nautilus.jpg" - } - - Dec 14 17:10:24.045: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . - Dec 14 17:10:24.045: INFO: update-demo-nautilus-sp2nw is verified up and running - STEP: scaling up the replication controller 12/14/22 17:10:24.045 - Dec 14 17:10:24.063: INFO: scanned /root for discovery docs: - Dec 14 17:10:24.063: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 scale rc update-demo-nautilus --replicas=2 --timeout=5m' - Dec 14 17:10:25.280: INFO: stderr: "" - Dec 14 17:10:25.280: INFO: stdout: "replicationcontroller/update-demo-nautilus scaled\n" - STEP: waiting for all containers in name=update-demo pods to come up. 12/14/22 17:10:25.28 - Dec 14 17:10:25.280: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' - Dec 14 17:10:25.456: INFO: stderr: "" - Dec 14 17:10:25.456: INFO: stdout: "update-demo-nautilus-nfpw2 update-demo-nautilus-sp2nw " - Dec 14 17:10:25.456: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-nfpw2 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' - Dec 14 17:10:25.658: INFO: stderr: "" - Dec 14 17:10:25.658: INFO: stdout: "" - Dec 14 17:10:25.659: INFO: update-demo-nautilus-nfpw2 is created but not running - Dec 14 17:10:30.659: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods -o template --template={{range.items}}{{.metadata.name}} {{end}} -l name=update-demo' - Dec 14 17:10:30.806: INFO: stderr: "" - Dec 14 17:10:30.806: INFO: stdout: "update-demo-nautilus-nfpw2 update-demo-nautilus-sp2nw " - Dec 14 17:10:30.807: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-nfpw2 -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' - Dec 14 17:10:31.011: INFO: stderr: "" - Dec 14 17:10:31.011: INFO: stdout: "true" - Dec 14 17:10:31.011: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-nfpw2 -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' - Dec 14 17:10:31.167: INFO: stderr: "" - Dec 14 17:10:31.167: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" - Dec 14 17:10:31.167: INFO: validating pod update-demo-nautilus-nfpw2 - Dec 14 17:10:31.182: INFO: got data: { - "image": "nautilus.jpg" - } - - Dec 14 17:10:31.182: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . - Dec 14 17:10:31.182: INFO: update-demo-nautilus-nfpw2 is verified up and running - Dec 14 17:10:31.182: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-sp2nw -o template --template={{if (exists . "status" "containerStatuses")}}{{range .status.containerStatuses}}{{if (and (eq .name "update-demo") (exists . "state" "running"))}}true{{end}}{{end}}{{end}}' - Dec 14 17:10:31.339: INFO: stderr: "" - Dec 14 17:10:31.339: INFO: stdout: "true" - Dec 14 17:10:31.340: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods update-demo-nautilus-sp2nw -o template --template={{if (exists . "spec" "containers")}}{{range .spec.containers}}{{if eq .name "update-demo"}}{{.image}}{{end}}{{end}}{{end}}' - Dec 14 17:10:31.454: INFO: stderr: "" - Dec 14 17:10:31.454: INFO: stdout: "registry.k8s.io/e2e-test-images/nautilus:1.5" - Dec 14 17:10:31.454: INFO: validating pod update-demo-nautilus-sp2nw - Dec 14 17:10:31.464: INFO: got data: { - "image": "nautilus.jpg" - } - - Dec 14 17:10:31.464: INFO: Unmarshalled json jpg/img => {nautilus.jpg} , expecting nautilus.jpg . - Dec 14 17:10:31.464: INFO: update-demo-nautilus-sp2nw is verified up and running - STEP: using delete to clean up resources 12/14/22 17:10:31.464 - Dec 14 17:10:31.464: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 delete --grace-period=0 --force -f -' - Dec 14 17:10:31.602: INFO: stderr: "Warning: Immediate deletion does not wait for confirmation that the running resource has been terminated. The resource may continue to run on the cluster indefinitely.\n" - Dec 14 17:10:31.602: INFO: stdout: "replicationcontroller \"update-demo-nautilus\" force deleted\n" - Dec 14 17:10:31.603: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get rc,svc -l name=update-demo --no-headers' - Dec 14 17:10:31.745: INFO: stderr: "No resources found in kubectl-3777 namespace.\n" - Dec 14 17:10:31.745: INFO: stdout: "" - Dec 14 17:10:31.745: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=kubectl-3777 get pods -l name=update-demo -o go-template={{ range .items }}{{ if not .metadata.deletionTimestamp }}{{ .metadata.name }}{{ "\n" }}{{ end }}{{ end }}' - Dec 14 17:10:32.009: INFO: stderr: "" - Dec 14 17:10:32.009: INFO: stdout: "" - [AfterEach] [sig-cli] Kubectl client - test/e2e/framework/framework.go:187 - Dec 14 17:10:32.010: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "kubectl-3777" for this suite. 12/14/22 17:10:32.017 - << End Captured GinkgoWriter Output +SSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-network] Services should be able to change the type from ClusterIP to ExternalName [Conformance] test/e2e/network/service.go:1481 [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:10:32.027 -Dec 14 17:10:32.027: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename services 12/14/22 17:10:32.03 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:10:32.058 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:10:32.06 +STEP: Creating a kubernetes client 02/25/23 11:42:03.012 +Feb 25 11:42:03.012: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename services 02/25/23 11:42:03.014 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:03.039 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:03.044 [BeforeEach] [sig-network] Services test/e2e/network/service.go:758 [It] should be able to change the type from ClusterIP to ExternalName [Conformance] test/e2e/network/service.go:1481 -STEP: creating a service clusterip-service with the type=ClusterIP in namespace services-5925 12/14/22 17:10:32.063 -STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service 12/14/22 17:10:32.121 -STEP: creating service externalsvc in namespace services-5925 12/14/22 17:10:32.122 -STEP: creating replication controller externalsvc in namespace services-5925 12/14/22 17:10:32.181 -I1214 17:10:32.207392 14 runners.go:193] Created replication controller with name: externalsvc, namespace: services-5925, replica count: 2 -I1214 17:10:35.258455 14 runners.go:193] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -STEP: changing the ClusterIP service to type=ExternalName 12/14/22 17:10:35.265 -Dec 14 17:10:35.286: INFO: Creating new exec pod -Dec 14 17:10:35.301: INFO: Waiting up to 5m0s for pod "execpod8d8cb" in namespace "services-5925" to be "running" -Dec 14 17:10:35.312: INFO: Pod "execpod8d8cb": Phase="Pending", Reason="", readiness=false. Elapsed: 10.752071ms -Dec 14 17:10:37.321: INFO: Pod "execpod8d8cb": Phase="Running", Reason="", readiness=true. Elapsed: 2.019415105s -Dec 14 17:10:37.321: INFO: Pod "execpod8d8cb" satisfied condition "running" -Dec 14 17:10:37.322: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5925 exec execpod8d8cb -- /bin/sh -x -c nslookup clusterip-service.services-5925.svc.cluster.local' -Dec 14 17:10:37.736: INFO: stderr: "+ nslookup clusterip-service.services-5925.svc.cluster.local\n" -Dec 14 17:10:37.736: INFO: stdout: "Server:\t\t10.233.0.10\nAddress:\t10.233.0.10#53\n\nclusterip-service.services-5925.svc.cluster.local\tcanonical name = externalsvc.services-5925.svc.cluster.local.\nName:\texternalsvc.services-5925.svc.cluster.local\nAddress: 10.233.18.178\n\n" -STEP: deleting ReplicationController externalsvc in namespace services-5925, will wait for the garbage collector to delete the pods 12/14/22 17:10:37.736 -Dec 14 17:10:37.811: INFO: Deleting ReplicationController externalsvc took: 17.624988ms -Dec 14 17:10:37.912: INFO: Terminating ReplicationController externalsvc pods took: 100.560222ms -Dec 14 17:10:41.644: INFO: Cleaning up the ClusterIP to ExternalName test service +STEP: creating a service clusterip-service with the type=ClusterIP in namespace services-6979 02/25/23 11:42:03.048 +STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service 02/25/23 11:42:03.064 +STEP: creating service externalsvc in namespace services-6979 02/25/23 11:42:03.064 +STEP: creating replication controller externalsvc in namespace services-6979 02/25/23 11:42:03.108 +I0225 11:42:03.127937 13 runners.go:193] Created replication controller with name: externalsvc, namespace: services-6979, replica count: 2 +I0225 11:42:06.187849 13 runners.go:193] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady +STEP: changing the ClusterIP service to type=ExternalName 02/25/23 11:42:06.192 +Feb 25 11:42:06.217: INFO: Creating new exec pod +Feb 25 11:42:06.236: INFO: Waiting up to 5m0s for pod "execpod6vb7h" in namespace "services-6979" to be "running" +Feb 25 11:42:06.250: INFO: Pod "execpod6vb7h": Phase="Pending", Reason="", readiness=false. Elapsed: 11.285101ms +Feb 25 11:42:08.256: INFO: Pod "execpod6vb7h": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017257832s +Feb 25 11:42:10.257: INFO: Pod "execpod6vb7h": Phase="Running", Reason="", readiness=true. Elapsed: 4.017557279s +Feb 25 11:42:10.257: INFO: Pod "execpod6vb7h" satisfied condition "running" +Feb 25 11:42:10.257: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6979 exec execpod6vb7h -- /bin/sh -x -c nslookup clusterip-service.services-6979.svc.cluster.local' +Feb 25 11:42:10.799: INFO: stderr: "+ nslookup clusterip-service.services-6979.svc.cluster.local\n" +Feb 25 11:42:10.799: INFO: stdout: "Server:\t\t10.233.0.10\nAddress:\t10.233.0.10#53\n\nclusterip-service.services-6979.svc.cluster.local\tcanonical name = externalsvc.services-6979.svc.cluster.local.\nName:\texternalsvc.services-6979.svc.cluster.local\nAddress: 10.233.5.163\n\n" +STEP: deleting ReplicationController externalsvc in namespace services-6979, will wait for the garbage collector to delete the pods 02/25/23 11:42:10.799 +Feb 25 11:42:10.869: INFO: Deleting ReplicationController externalsvc took: 12.116721ms +Feb 25 11:42:10.972: INFO: Terminating ReplicationController externalsvc pods took: 103.449491ms +Feb 25 11:42:13.632: INFO: Cleaning up the ClusterIP to ExternalName test service [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 -Dec 14 17:10:41.659: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "services-5925" for this suite. 12/14/22 17:10:41.666 +Feb 25 11:42:13.683: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "services-6979" for this suite. 02/25/23 11:42:13.702 [AfterEach] [sig-network] Services test/e2e/network/service.go:762 -{"msg":"PASSED [sig-network] Services should be able to change the type from ClusterIP to ExternalName [Conformance]","completed":354,"skipped":6562,"failed":0} +{"msg":"PASSED [sig-network] Services should be able to change the type from ClusterIP to ExternalName [Conformance]","completed":351,"skipped":6537,"failed":0} ------------------------------ -• [SLOW TEST] [9.649 seconds] +• [SLOW TEST] [10.705 seconds] [sig-network] Services test/e2e/network/common/framework.go:23 should be able to change the type from ClusterIP to ExternalName [Conformance] @@ -30669,1483 +32188,912 @@ test/e2e/network/common/framework.go:23 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-network] Services test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:10:32.027 - Dec 14 17:10:32.027: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename services 12/14/22 17:10:32.03 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:10:32.058 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:10:32.06 + STEP: Creating a kubernetes client 02/25/23 11:42:03.012 + Feb 25 11:42:03.012: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename services 02/25/23 11:42:03.014 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:03.039 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:03.044 [BeforeEach] [sig-network] Services test/e2e/network/service.go:758 [It] should be able to change the type from ClusterIP to ExternalName [Conformance] test/e2e/network/service.go:1481 - STEP: creating a service clusterip-service with the type=ClusterIP in namespace services-5925 12/14/22 17:10:32.063 - STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service 12/14/22 17:10:32.121 - STEP: creating service externalsvc in namespace services-5925 12/14/22 17:10:32.122 - STEP: creating replication controller externalsvc in namespace services-5925 12/14/22 17:10:32.181 - I1214 17:10:32.207392 14 runners.go:193] Created replication controller with name: externalsvc, namespace: services-5925, replica count: 2 - I1214 17:10:35.258455 14 runners.go:193] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - STEP: changing the ClusterIP service to type=ExternalName 12/14/22 17:10:35.265 - Dec 14 17:10:35.286: INFO: Creating new exec pod - Dec 14 17:10:35.301: INFO: Waiting up to 5m0s for pod "execpod8d8cb" in namespace "services-5925" to be "running" - Dec 14 17:10:35.312: INFO: Pod "execpod8d8cb": Phase="Pending", Reason="", readiness=false. Elapsed: 10.752071ms - Dec 14 17:10:37.321: INFO: Pod "execpod8d8cb": Phase="Running", Reason="", readiness=true. Elapsed: 2.019415105s - Dec 14 17:10:37.321: INFO: Pod "execpod8d8cb" satisfied condition "running" - Dec 14 17:10:37.322: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-3507870282 --namespace=services-5925 exec execpod8d8cb -- /bin/sh -x -c nslookup clusterip-service.services-5925.svc.cluster.local' - Dec 14 17:10:37.736: INFO: stderr: "+ nslookup clusterip-service.services-5925.svc.cluster.local\n" - Dec 14 17:10:37.736: INFO: stdout: "Server:\t\t10.233.0.10\nAddress:\t10.233.0.10#53\n\nclusterip-service.services-5925.svc.cluster.local\tcanonical name = externalsvc.services-5925.svc.cluster.local.\nName:\texternalsvc.services-5925.svc.cluster.local\nAddress: 10.233.18.178\n\n" - STEP: deleting ReplicationController externalsvc in namespace services-5925, will wait for the garbage collector to delete the pods 12/14/22 17:10:37.736 - Dec 14 17:10:37.811: INFO: Deleting ReplicationController externalsvc took: 17.624988ms - Dec 14 17:10:37.912: INFO: Terminating ReplicationController externalsvc pods took: 100.560222ms - Dec 14 17:10:41.644: INFO: Cleaning up the ClusterIP to ExternalName test service + STEP: creating a service clusterip-service with the type=ClusterIP in namespace services-6979 02/25/23 11:42:03.048 + STEP: Creating active service to test reachability when its FQDN is referred as externalName for another service 02/25/23 11:42:03.064 + STEP: creating service externalsvc in namespace services-6979 02/25/23 11:42:03.064 + STEP: creating replication controller externalsvc in namespace services-6979 02/25/23 11:42:03.108 + I0225 11:42:03.127937 13 runners.go:193] Created replication controller with name: externalsvc, namespace: services-6979, replica count: 2 + I0225 11:42:06.187849 13 runners.go:193] externalsvc Pods: 2 out of 2 created, 2 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady + STEP: changing the ClusterIP service to type=ExternalName 02/25/23 11:42:06.192 + Feb 25 11:42:06.217: INFO: Creating new exec pod + Feb 25 11:42:06.236: INFO: Waiting up to 5m0s for pod "execpod6vb7h" in namespace "services-6979" to be "running" + Feb 25 11:42:06.250: INFO: Pod "execpod6vb7h": Phase="Pending", Reason="", readiness=false. Elapsed: 11.285101ms + Feb 25 11:42:08.256: INFO: Pod "execpod6vb7h": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017257832s + Feb 25 11:42:10.257: INFO: Pod "execpod6vb7h": Phase="Running", Reason="", readiness=true. Elapsed: 4.017557279s + Feb 25 11:42:10.257: INFO: Pod "execpod6vb7h" satisfied condition "running" + Feb 25 11:42:10.257: INFO: Running '/usr/local/bin/kubectl --kubeconfig=/tmp/kubeconfig-2917274343 --namespace=services-6979 exec execpod6vb7h -- /bin/sh -x -c nslookup clusterip-service.services-6979.svc.cluster.local' + Feb 25 11:42:10.799: INFO: stderr: "+ nslookup clusterip-service.services-6979.svc.cluster.local\n" + Feb 25 11:42:10.799: INFO: stdout: "Server:\t\t10.233.0.10\nAddress:\t10.233.0.10#53\n\nclusterip-service.services-6979.svc.cluster.local\tcanonical name = externalsvc.services-6979.svc.cluster.local.\nName:\texternalsvc.services-6979.svc.cluster.local\nAddress: 10.233.5.163\n\n" + STEP: deleting ReplicationController externalsvc in namespace services-6979, will wait for the garbage collector to delete the pods 02/25/23 11:42:10.799 + Feb 25 11:42:10.869: INFO: Deleting ReplicationController externalsvc took: 12.116721ms + Feb 25 11:42:10.972: INFO: Terminating ReplicationController externalsvc pods took: 103.449491ms + Feb 25 11:42:13.632: INFO: Cleaning up the ClusterIP to ExternalName test service [AfterEach] [sig-network] Services test/e2e/framework/framework.go:187 - Dec 14 17:10:41.659: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "services-5925" for this suite. 12/14/22 17:10:41.666 + Feb 25 11:42:13.683: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "services-6979" for this suite. 02/25/23 11:42:13.702 [AfterEach] [sig-network] Services test/e2e/network/service.go:762 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSS +------------------------------ +[sig-storage] Downward API volume + should provide container's cpu request [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:220 +[BeforeEach] [sig-storage] Downward API volume + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 11:42:13.718 +Feb 25 11:42:13.718: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename downward-api 02/25/23 11:42:13.721 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:13.791 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:13.795 +[BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 +[It] should provide container's cpu request [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:220 +STEP: Creating a pod to test downward API volume plugin 02/25/23 11:42:13.812 +Feb 25 11:42:13.830: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201" in namespace "downward-api-1198" to be "Succeeded or Failed" +Feb 25 11:42:13.839: INFO: Pod "downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201": Phase="Pending", Reason="", readiness=false. Elapsed: 8.874373ms +Feb 25 11:42:15.847: INFO: Pod "downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017173489s +Feb 25 11:42:17.846: INFO: Pod "downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201": Phase="Running", Reason="", readiness=false. Elapsed: 4.016166397s +Feb 25 11:42:19.849: INFO: Pod "downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.018991523s +STEP: Saw pod success 02/25/23 11:42:19.849 +Feb 25 11:42:19.849: INFO: Pod "downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201" satisfied condition "Succeeded or Failed" +Feb 25 11:42:19.853: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201 container client-container: +STEP: delete the pod 02/25/23 11:42:20.16 +Feb 25 11:42:20.179: INFO: Waiting for pod downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201 to disappear +Feb 25 11:42:20.183: INFO: Pod downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201 no longer exists +[AfterEach] [sig-storage] Downward API volume + test/e2e/framework/framework.go:187 +Feb 25 11:42:20.184: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "downward-api-1198" for this suite. 02/25/23 11:42:20.191 +{"msg":"PASSED [sig-storage] Downward API volume should provide container's cpu request [NodeConformance] [Conformance]","completed":352,"skipped":6551,"failed":0} +------------------------------ +• [SLOW TEST] [6.484 seconds] +[sig-storage] Downward API volume +test/e2e/common/storage/framework.go:23 + should provide container's cpu request [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:220 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-storage] Downward API volume + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 11:42:13.718 + Feb 25 11:42:13.718: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename downward-api 02/25/23 11:42:13.721 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:13.791 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:13.795 + [BeforeEach] [sig-storage] Downward API volume + test/e2e/common/storage/downwardapi_volume.go:43 + [It] should provide container's cpu request [NodeConformance] [Conformance] + test/e2e/common/storage/downwardapi_volume.go:220 + STEP: Creating a pod to test downward API volume plugin 02/25/23 11:42:13.812 + Feb 25 11:42:13.830: INFO: Waiting up to 5m0s for pod "downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201" in namespace "downward-api-1198" to be "Succeeded or Failed" + Feb 25 11:42:13.839: INFO: Pod "downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201": Phase="Pending", Reason="", readiness=false. Elapsed: 8.874373ms + Feb 25 11:42:15.847: INFO: Pod "downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201": Phase="Pending", Reason="", readiness=false. Elapsed: 2.017173489s + Feb 25 11:42:17.846: INFO: Pod "downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201": Phase="Running", Reason="", readiness=false. Elapsed: 4.016166397s + Feb 25 11:42:19.849: INFO: Pod "downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.018991523s + STEP: Saw pod success 02/25/23 11:42:19.849 + Feb 25 11:42:19.849: INFO: Pod "downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201" satisfied condition "Succeeded or Failed" + Feb 25 11:42:19.853: INFO: Trying to get logs from node diek3dua9gay-3 pod downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201 container client-container: + STEP: delete the pod 02/25/23 11:42:20.16 + Feb 25 11:42:20.179: INFO: Waiting for pod downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201 to disappear + Feb 25 11:42:20.183: INFO: Pod downwardapi-volume-d6dc971c-e9b9-4a50-907b-e327919ee201 no longer exists + [AfterEach] [sig-storage] Downward API volume + test/e2e/framework/framework.go:187 + Feb 25 11:42:20.184: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "downward-api-1198" for this suite. 02/25/23 11:42:20.191 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS ------------------------------ [sig-storage] ConfigMap - should be consumable from pods in volume as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:73 + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:88 [BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:10:41.692 -Dec 14 17:10:41.692: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename configmap 12/14/22 17:10:41.694 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:10:41.728 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:10:41.734 -[It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:73 -STEP: Creating configMap with name configmap-test-volume-6c88dfc5-d1af-4a2b-b1c5-923910c0d9aa 12/14/22 17:10:41.74 -STEP: Creating a pod to test consume configMaps 12/14/22 17:10:41.748 -Dec 14 17:10:41.767: INFO: Waiting up to 5m0s for pod "pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11" in namespace "configmap-9791" to be "Succeeded or Failed" -Dec 14 17:10:41.794: INFO: Pod "pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11": Phase="Pending", Reason="", readiness=false. Elapsed: 26.644967ms -Dec 14 17:10:43.801: INFO: Pod "pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033359998s -Dec 14 17:10:45.813: INFO: Pod "pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.046101081s -STEP: Saw pod success 12/14/22 17:10:45.814 -Dec 14 17:10:45.814: INFO: Pod "pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11" satisfied condition "Succeeded or Failed" -Dec 14 17:10:45.839: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11 container agnhost-container: -STEP: delete the pod 12/14/22 17:10:45.995 -Dec 14 17:10:46.014: INFO: Waiting for pod pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11 to disappear -Dec 14 17:10:46.019: INFO: Pod pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11 no longer exists +STEP: Creating a kubernetes client 02/25/23 11:42:20.211 +Feb 25 11:42:20.212: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename configmap 02/25/23 11:42:20.214 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:20.239 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:20.243 +[It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:88 +STEP: Creating configMap with name configmap-test-volume-map-0cd66fe5-f77c-47c0-89f0-8bf0c9ac47c5 02/25/23 11:42:20.247 +STEP: Creating a pod to test consume configMaps 02/25/23 11:42:20.255 +Feb 25 11:42:20.273: INFO: Waiting up to 5m0s for pod "pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace" in namespace "configmap-8984" to be "Succeeded or Failed" +Feb 25 11:42:20.279: INFO: Pod "pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace": Phase="Pending", Reason="", readiness=false. Elapsed: 5.634079ms +Feb 25 11:42:22.287: INFO: Pod "pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01362202s +Feb 25 11:42:24.286: INFO: Pod "pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012528734s +Feb 25 11:42:26.286: INFO: Pod "pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.013259182s +STEP: Saw pod success 02/25/23 11:42:26.286 +Feb 25 11:42:26.286: INFO: Pod "pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace" satisfied condition "Succeeded or Failed" +Feb 25 11:42:26.291: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace container agnhost-container: +STEP: delete the pod 02/25/23 11:42:26.301 +Feb 25 11:42:26.316: INFO: Waiting for pod pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace to disappear +Feb 25 11:42:26.320: INFO: Pod pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace no longer exists [AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 -Dec 14 17:10:46.020: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "configmap-9791" for this suite. 12/14/22 17:10:46.031 -{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume as non-root [NodeConformance] [Conformance]","completed":355,"skipped":6617,"failed":0} +Feb 25 11:42:26.320: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "configmap-8984" for this suite. 02/25/23 11:42:26.327 +{"msg":"PASSED [sig-storage] ConfigMap should be consumable from pods in volume with mappings [NodeConformance] [Conformance]","completed":353,"skipped":6605,"failed":0} ------------------------------ -• [4.350 seconds] +• [SLOW TEST] [6.128 seconds] [sig-storage] ConfigMap test/e2e/common/storage/framework.go:23 - should be consumable from pods in volume as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:73 + should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:88 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:10:41.692 - Dec 14 17:10:41.692: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename configmap 12/14/22 17:10:41.694 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:10:41.728 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:10:41.734 - [It] should be consumable from pods in volume as non-root [NodeConformance] [Conformance] - test/e2e/common/storage/configmap_volume.go:73 - STEP: Creating configMap with name configmap-test-volume-6c88dfc5-d1af-4a2b-b1c5-923910c0d9aa 12/14/22 17:10:41.74 - STEP: Creating a pod to test consume configMaps 12/14/22 17:10:41.748 - Dec 14 17:10:41.767: INFO: Waiting up to 5m0s for pod "pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11" in namespace "configmap-9791" to be "Succeeded or Failed" - Dec 14 17:10:41.794: INFO: Pod "pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11": Phase="Pending", Reason="", readiness=false. Elapsed: 26.644967ms - Dec 14 17:10:43.801: INFO: Pod "pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11": Phase="Pending", Reason="", readiness=false. Elapsed: 2.033359998s - Dec 14 17:10:45.813: INFO: Pod "pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11": Phase="Succeeded", Reason="", readiness=false. Elapsed: 4.046101081s - STEP: Saw pod success 12/14/22 17:10:45.814 - Dec 14 17:10:45.814: INFO: Pod "pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11" satisfied condition "Succeeded or Failed" - Dec 14 17:10:45.839: INFO: Trying to get logs from node kiedai4iakae-3 pod pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11 container agnhost-container: - STEP: delete the pod 12/14/22 17:10:45.995 - Dec 14 17:10:46.014: INFO: Waiting for pod pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11 to disappear - Dec 14 17:10:46.019: INFO: Pod pod-configmaps-ef42da86-b591-45d4-adfa-bbb9add11c11 no longer exists + STEP: Creating a kubernetes client 02/25/23 11:42:20.211 + Feb 25 11:42:20.212: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename configmap 02/25/23 11:42:20.214 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:20.239 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:20.243 + [It] should be consumable from pods in volume with mappings [NodeConformance] [Conformance] + test/e2e/common/storage/configmap_volume.go:88 + STEP: Creating configMap with name configmap-test-volume-map-0cd66fe5-f77c-47c0-89f0-8bf0c9ac47c5 02/25/23 11:42:20.247 + STEP: Creating a pod to test consume configMaps 02/25/23 11:42:20.255 + Feb 25 11:42:20.273: INFO: Waiting up to 5m0s for pod "pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace" in namespace "configmap-8984" to be "Succeeded or Failed" + Feb 25 11:42:20.279: INFO: Pod "pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace": Phase="Pending", Reason="", readiness=false. Elapsed: 5.634079ms + Feb 25 11:42:22.287: INFO: Pod "pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace": Phase="Pending", Reason="", readiness=false. Elapsed: 2.01362202s + Feb 25 11:42:24.286: INFO: Pod "pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012528734s + Feb 25 11:42:26.286: INFO: Pod "pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.013259182s + STEP: Saw pod success 02/25/23 11:42:26.286 + Feb 25 11:42:26.286: INFO: Pod "pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace" satisfied condition "Succeeded or Failed" + Feb 25 11:42:26.291: INFO: Trying to get logs from node diek3dua9gay-3 pod pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace container agnhost-container: + STEP: delete the pod 02/25/23 11:42:26.301 + Feb 25 11:42:26.316: INFO: Waiting for pod pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace to disappear + Feb 25 11:42:26.320: INFO: Pod pod-configmaps-0df41e5c-adb7-480f-8945-a4795c19eace no longer exists [AfterEach] [sig-storage] ConfigMap test/e2e/framework/framework.go:187 - Dec 14 17:10:46.020: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "configmap-9791" for this suite. 12/14/22 17:10:46.031 + Feb 25 11:42:26.320: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "configmap-8984" for this suite. 02/25/23 11:42:26.327 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSS ------------------------------ -[sig-storage] Secrets - should be immutable if `immutable` field is set [Conformance] - test/e2e/common/storage/secrets_volume.go:385 -[BeforeEach] [sig-storage] Secrets +[sig-apps] DisruptionController + should create a PodDisruptionBudget [Conformance] + test/e2e/apps/disruption.go:107 +[BeforeEach] [sig-apps] DisruptionController + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 11:42:26.343 +Feb 25 11:42:26.344: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename disruption 02/25/23 11:42:26.346 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:26.375 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:26.379 +[BeforeEach] [sig-apps] DisruptionController + test/e2e/apps/disruption.go:71 +[It] should create a PodDisruptionBudget [Conformance] + test/e2e/apps/disruption.go:107 +STEP: creating the pdb 02/25/23 11:42:26.383 +STEP: Waiting for the pdb to be processed 02/25/23 11:42:26.396 +STEP: updating the pdb 02/25/23 11:42:26.404 +STEP: Waiting for the pdb to be processed 02/25/23 11:42:26.422 +STEP: patching the pdb 02/25/23 11:42:28.438 +STEP: Waiting for the pdb to be processed 02/25/23 11:42:28.452 +STEP: Waiting for the pdb to be deleted 02/25/23 11:42:30.477 +[AfterEach] [sig-apps] DisruptionController + test/e2e/framework/framework.go:187 +Feb 25 11:42:30.484: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "disruption-5932" for this suite. 02/25/23 11:42:30.493 +{"msg":"PASSED [sig-apps] DisruptionController should create a PodDisruptionBudget [Conformance]","completed":354,"skipped":6617,"failed":0} +------------------------------ +• [4.162 seconds] +[sig-apps] DisruptionController +test/e2e/apps/framework.go:23 + should create a PodDisruptionBudget [Conformance] + test/e2e/apps/disruption.go:107 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-apps] DisruptionController + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 11:42:26.343 + Feb 25 11:42:26.344: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename disruption 02/25/23 11:42:26.346 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:26.375 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:26.379 + [BeforeEach] [sig-apps] DisruptionController + test/e2e/apps/disruption.go:71 + [It] should create a PodDisruptionBudget [Conformance] + test/e2e/apps/disruption.go:107 + STEP: creating the pdb 02/25/23 11:42:26.383 + STEP: Waiting for the pdb to be processed 02/25/23 11:42:26.396 + STEP: updating the pdb 02/25/23 11:42:26.404 + STEP: Waiting for the pdb to be processed 02/25/23 11:42:26.422 + STEP: patching the pdb 02/25/23 11:42:28.438 + STEP: Waiting for the pdb to be processed 02/25/23 11:42:28.452 + STEP: Waiting for the pdb to be deleted 02/25/23 11:42:30.477 + [AfterEach] [sig-apps] DisruptionController + test/e2e/framework/framework.go:187 + Feb 25 11:42:30.484: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "disruption-5932" for this suite. 02/25/23 11:42:30.493 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSSSSSSSSSSSS +------------------------------ +[sig-node] Secrets + should fail to create secret due to empty secret key [Conformance] + test/e2e/common/node/secrets.go:139 +[BeforeEach] [sig-node] Secrets + test/e2e/framework/framework.go:186 +STEP: Creating a kubernetes client 02/25/23 11:42:30.517 +Feb 25 11:42:30.517: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename secrets 02/25/23 11:42:30.519 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:30.549 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:30.555 +[It] should fail to create secret due to empty secret key [Conformance] + test/e2e/common/node/secrets.go:139 +STEP: Creating projection with secret that has name secret-emptykey-test-5bbccd8c-01a2-415e-89b1-45f9a9e59b4c 02/25/23 11:42:30.559 +[AfterEach] [sig-node] Secrets + test/e2e/framework/framework.go:187 +Feb 25 11:42:30.562: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "secrets-4684" for this suite. 02/25/23 11:42:30.569 +{"msg":"PASSED [sig-node] Secrets should fail to create secret due to empty secret key [Conformance]","completed":355,"skipped":6636,"failed":0} +------------------------------ +• [0.062 seconds] +[sig-node] Secrets +test/e2e/common/node/framework.go:23 + should fail to create secret due to empty secret key [Conformance] + test/e2e/common/node/secrets.go:139 + + Begin Captured GinkgoWriter Output >> + [BeforeEach] [sig-node] Secrets + test/e2e/framework/framework.go:186 + STEP: Creating a kubernetes client 02/25/23 11:42:30.517 + Feb 25 11:42:30.517: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename secrets 02/25/23 11:42:30.519 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:30.549 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:30.555 + [It] should fail to create secret due to empty secret key [Conformance] + test/e2e/common/node/secrets.go:139 + STEP: Creating projection with secret that has name secret-emptykey-test-5bbccd8c-01a2-415e-89b1-45f9a9e59b4c 02/25/23 11:42:30.559 + [AfterEach] [sig-node] Secrets + test/e2e/framework/framework.go:187 + Feb 25 11:42:30.562: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "secrets-4684" for this suite. 02/25/23 11:42:30.569 + << End Captured GinkgoWriter Output +------------------------------ +SSSSSSSSS +------------------------------ +[sig-node] Security Context When creating a pod with privileged + should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:527 +[BeforeEach] [sig-node] Security Context test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:10:46.047 -Dec 14 17:10:46.047: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename secrets 12/14/22 17:10:46.052 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:10:46.082 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:10:46.086 -[It] should be immutable if `immutable` field is set [Conformance] - test/e2e/common/storage/secrets_volume.go:385 -[AfterEach] [sig-storage] Secrets +STEP: Creating a kubernetes client 02/25/23 11:42:30.58 +Feb 25 11:42:30.580: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename security-context-test 02/25/23 11:42:30.585 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:30.627 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:30.633 +[BeforeEach] [sig-node] Security Context + test/e2e/common/node/security_context.go:49 +[It] should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:527 +Feb 25 11:42:30.655: INFO: Waiting up to 5m0s for pod "busybox-privileged-false-79a85c2b-73b9-47af-ac37-e0e46db6d5ac" in namespace "security-context-test-4187" to be "Succeeded or Failed" +Feb 25 11:42:30.660: INFO: Pod "busybox-privileged-false-79a85c2b-73b9-47af-ac37-e0e46db6d5ac": Phase="Pending", Reason="", readiness=false. Elapsed: 5.014525ms +Feb 25 11:42:32.667: INFO: Pod "busybox-privileged-false-79a85c2b-73b9-47af-ac37-e0e46db6d5ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012023375s +Feb 25 11:42:34.667: INFO: Pod "busybox-privileged-false-79a85c2b-73b9-47af-ac37-e0e46db6d5ac": Phase="Pending", Reason="", readiness=false. Elapsed: 4.01227006s +Feb 25 11:42:36.667: INFO: Pod "busybox-privileged-false-79a85c2b-73b9-47af-ac37-e0e46db6d5ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.011702974s +Feb 25 11:42:36.667: INFO: Pod "busybox-privileged-false-79a85c2b-73b9-47af-ac37-e0e46db6d5ac" satisfied condition "Succeeded or Failed" +Feb 25 11:42:36.676: INFO: Got logs for pod "busybox-privileged-false-79a85c2b-73b9-47af-ac37-e0e46db6d5ac": "ip: RTNETLINK answers: Operation not permitted\n" +[AfterEach] [sig-node] Security Context test/e2e/framework/framework.go:187 -Dec 14 17:10:46.174: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "secrets-8868" for this suite. 12/14/22 17:10:46.18 -{"msg":"PASSED [sig-storage] Secrets should be immutable if `immutable` field is set [Conformance]","completed":356,"skipped":6637,"failed":0} +Feb 25 11:42:36.677: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "security-context-test-4187" for this suite. 02/25/23 11:42:36.698 +{"msg":"PASSED [sig-node] Security Context When creating a pod with privileged should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance]","completed":356,"skipped":6645,"failed":0} ------------------------------ -• [0.142 seconds] -[sig-storage] Secrets -test/e2e/common/storage/framework.go:23 - should be immutable if `immutable` field is set [Conformance] - test/e2e/common/storage/secrets_volume.go:385 +• [SLOW TEST] [6.133 seconds] +[sig-node] Security Context +test/e2e/common/node/framework.go:23 + When creating a pod with privileged + test/e2e/common/node/security_context.go:490 + should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:527 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] Secrets + [BeforeEach] [sig-node] Security Context test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:10:46.047 - Dec 14 17:10:46.047: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename secrets 12/14/22 17:10:46.052 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:10:46.082 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:10:46.086 - [It] should be immutable if `immutable` field is set [Conformance] - test/e2e/common/storage/secrets_volume.go:385 - [AfterEach] [sig-storage] Secrets + STEP: Creating a kubernetes client 02/25/23 11:42:30.58 + Feb 25 11:42:30.580: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename security-context-test 02/25/23 11:42:30.585 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:30.627 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:30.633 + [BeforeEach] [sig-node] Security Context + test/e2e/common/node/security_context.go:49 + [It] should run the container as unprivileged when false [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/node/security_context.go:527 + Feb 25 11:42:30.655: INFO: Waiting up to 5m0s for pod "busybox-privileged-false-79a85c2b-73b9-47af-ac37-e0e46db6d5ac" in namespace "security-context-test-4187" to be "Succeeded or Failed" + Feb 25 11:42:30.660: INFO: Pod "busybox-privileged-false-79a85c2b-73b9-47af-ac37-e0e46db6d5ac": Phase="Pending", Reason="", readiness=false. Elapsed: 5.014525ms + Feb 25 11:42:32.667: INFO: Pod "busybox-privileged-false-79a85c2b-73b9-47af-ac37-e0e46db6d5ac": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012023375s + Feb 25 11:42:34.667: INFO: Pod "busybox-privileged-false-79a85c2b-73b9-47af-ac37-e0e46db6d5ac": Phase="Pending", Reason="", readiness=false. Elapsed: 4.01227006s + Feb 25 11:42:36.667: INFO: Pod "busybox-privileged-false-79a85c2b-73b9-47af-ac37-e0e46db6d5ac": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.011702974s + Feb 25 11:42:36.667: INFO: Pod "busybox-privileged-false-79a85c2b-73b9-47af-ac37-e0e46db6d5ac" satisfied condition "Succeeded or Failed" + Feb 25 11:42:36.676: INFO: Got logs for pod "busybox-privileged-false-79a85c2b-73b9-47af-ac37-e0e46db6d5ac": "ip: RTNETLINK answers: Operation not permitted\n" + [AfterEach] [sig-node] Security Context test/e2e/framework/framework.go:187 - Dec 14 17:10:46.174: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "secrets-8868" for this suite. 12/14/22 17:10:46.18 + Feb 25 11:42:36.677: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "security-context-test-4187" for this suite. 02/25/23 11:42:36.698 << End Captured GinkgoWriter Output ------------------------------ -SSSS +SSSSSSS ------------------------------ -[sig-network] EndpointSlice - should have Endpoints and EndpointSlices pointing to API Server [Conformance] - test/e2e/network/endpointslice.go:65 -[BeforeEach] [sig-network] EndpointSlice +[sig-network] Networking Granular Checks: Pods + should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:105 +[BeforeEach] [sig-network] Networking test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:10:46.189 -Dec 14 17:10:46.189: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename endpointslice 12/14/22 17:10:46.198 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:10:46.223 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:10:46.226 -[BeforeEach] [sig-network] EndpointSlice - test/e2e/network/endpointslice.go:51 -[It] should have Endpoints and EndpointSlices pointing to API Server [Conformance] - test/e2e/network/endpointslice.go:65 -Dec 14 17:10:46.239: INFO: Endpoints addresses: [192.168.121.197 192.168.121.249] , ports: [6443] -Dec 14 17:10:46.239: INFO: EndpointSlices addresses: [192.168.121.197 192.168.121.249] , ports: [6443] -[AfterEach] [sig-network] EndpointSlice +STEP: Creating a kubernetes client 02/25/23 11:42:36.726 +Feb 25 11:42:36.727: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename pod-network-test 02/25/23 11:42:36.733 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:36.759 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:36.764 +[It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:105 +STEP: Performing setup for networking test in namespace pod-network-test-627 02/25/23 11:42:36.769 +STEP: creating a selector 02/25/23 11:42:36.769 +STEP: Creating the service pods in kubernetes 02/25/23 11:42:36.77 +Feb 25 11:42:36.770: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable +Feb 25 11:42:36.822: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-627" to be "running and ready" +Feb 25 11:42:36.828: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 6.411857ms +Feb 25 11:42:36.828: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) +Feb 25 11:42:38.834: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 2.012230024s +Feb 25 11:42:38.834: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:42:40.895: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.073195665s +Feb 25 11:42:40.895: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:42:42.835: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.01283904s +Feb 25 11:42:42.835: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:42:44.834: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.012754701s +Feb 25 11:42:44.835: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:42:46.836: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.014398809s +Feb 25 11:42:46.836: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:42:48.835: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 12.013032378s +Feb 25 11:42:48.835: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:42:50.836: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 14.014072923s +Feb 25 11:42:50.836: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:42:52.837: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 16.015405894s +Feb 25 11:42:52.837: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:42:54.837: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 18.015591357s +Feb 25 11:42:54.837: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:42:56.838: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 20.015848659s +Feb 25 11:42:56.838: INFO: The phase of Pod netserver-0 is Running (Ready = false) +Feb 25 11:42:58.835: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 22.013461164s +Feb 25 11:42:58.835: INFO: The phase of Pod netserver-0 is Running (Ready = true) +Feb 25 11:42:58.835: INFO: Pod "netserver-0" satisfied condition "running and ready" +Feb 25 11:42:58.840: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-627" to be "running and ready" +Feb 25 11:42:58.846: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 5.390314ms +Feb 25 11:42:58.847: INFO: The phase of Pod netserver-1 is Running (Ready = true) +Feb 25 11:42:58.848: INFO: Pod "netserver-1" satisfied condition "running and ready" +Feb 25 11:42:58.854: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-627" to be "running and ready" +Feb 25 11:42:58.861: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 6.437109ms +Feb 25 11:42:58.861: INFO: The phase of Pod netserver-2 is Running (Ready = true) +Feb 25 11:42:58.861: INFO: Pod "netserver-2" satisfied condition "running and ready" +STEP: Creating test pods 02/25/23 11:42:58.866 +Feb 25 11:42:58.895: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-627" to be "running" +Feb 25 11:42:58.917: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 22.657371ms +Feb 25 11:43:00.930: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034823785s +Feb 25 11:43:02.929: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.034316993s +Feb 25 11:43:02.929: INFO: Pod "test-container-pod" satisfied condition "running" +Feb 25 11:43:02.934: INFO: Waiting up to 5m0s for pod "host-test-container-pod" in namespace "pod-network-test-627" to be "running" +Feb 25 11:43:02.956: INFO: Pod "host-test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 21.329021ms +Feb 25 11:43:02.956: INFO: Pod "host-test-container-pod" satisfied condition "running" +Feb 25 11:43:02.962: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 +Feb 25 11:43:02.962: INFO: Going to poll 10.233.65.137 on port 8083 at least 0 times, with a maximum of 39 tries before failing +Feb 25 11:43:02.975: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.233.65.137:8083/hostName | grep -v '^\s*$'] Namespace:pod-network-test-627 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 11:43:02.975: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 11:43:02.976: INFO: ExecWithOptions: Clientset creation +Feb 25 11:43:02.977: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-627/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+--max-time+15+--connect-timeout+1+http%3A%2F%2F10.233.65.137%3A8083%2FhostName+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) +Feb 25 11:43:03.184: INFO: Found all 1 expected endpoints: [netserver-0] +Feb 25 11:43:03.184: INFO: Going to poll 10.233.64.174 on port 8083 at least 0 times, with a maximum of 39 tries before failing +Feb 25 11:43:03.191: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.233.64.174:8083/hostName | grep -v '^\s*$'] Namespace:pod-network-test-627 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 11:43:03.191: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 11:43:03.192: INFO: ExecWithOptions: Clientset creation +Feb 25 11:43:03.192: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-627/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+--max-time+15+--connect-timeout+1+http%3A%2F%2F10.233.64.174%3A8083%2FhostName+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) +Feb 25 11:43:03.328: INFO: Found all 1 expected endpoints: [netserver-1] +Feb 25 11:43:03.328: INFO: Going to poll 10.233.66.35 on port 8083 at least 0 times, with a maximum of 39 tries before failing +Feb 25 11:43:03.333: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.233.66.35:8083/hostName | grep -v '^\s*$'] Namespace:pod-network-test-627 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} +Feb 25 11:43:03.334: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +Feb 25 11:43:03.335: INFO: ExecWithOptions: Clientset creation +Feb 25 11:43:03.335: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-627/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+--max-time+15+--connect-timeout+1+http%3A%2F%2F10.233.66.35%3A8083%2FhostName+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) +Feb 25 11:43:03.417: INFO: Found all 1 expected endpoints: [netserver-2] +[AfterEach] [sig-network] Networking test/e2e/framework/framework.go:187 -Dec 14 17:10:46.239: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "endpointslice-5410" for this suite. 12/14/22 17:10:46.245 -{"msg":"PASSED [sig-network] EndpointSlice should have Endpoints and EndpointSlices pointing to API Server [Conformance]","completed":357,"skipped":6641,"failed":0} +Feb 25 11:43:03.417: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "pod-network-test-627" for this suite. 02/25/23 11:43:03.427 +{"msg":"PASSED [sig-network] Networking Granular Checks: Pods should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance]","completed":357,"skipped":6652,"failed":0} ------------------------------ -• [0.063 seconds] -[sig-network] EndpointSlice -test/e2e/network/common/framework.go:23 - should have Endpoints and EndpointSlices pointing to API Server [Conformance] - test/e2e/network/endpointslice.go:65 +• [SLOW TEST] [26.712 seconds] +[sig-network] Networking +test/e2e/common/network/framework.go:23 + Granular Checks: Pods + test/e2e/common/network/networking.go:32 + should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:105 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] EndpointSlice + [BeforeEach] [sig-network] Networking test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:10:46.189 - Dec 14 17:10:46.189: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename endpointslice 12/14/22 17:10:46.198 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:10:46.223 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:10:46.226 - [BeforeEach] [sig-network] EndpointSlice - test/e2e/network/endpointslice.go:51 - [It] should have Endpoints and EndpointSlices pointing to API Server [Conformance] - test/e2e/network/endpointslice.go:65 - Dec 14 17:10:46.239: INFO: Endpoints addresses: [192.168.121.197 192.168.121.249] , ports: [6443] - Dec 14 17:10:46.239: INFO: EndpointSlices addresses: [192.168.121.197 192.168.121.249] , ports: [6443] - [AfterEach] [sig-network] EndpointSlice + STEP: Creating a kubernetes client 02/25/23 11:42:36.726 + Feb 25 11:42:36.727: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename pod-network-test 02/25/23 11:42:36.733 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:42:36.759 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:42:36.764 + [It] should function for node-pod communication: http [LinuxOnly] [NodeConformance] [Conformance] + test/e2e/common/network/networking.go:105 + STEP: Performing setup for networking test in namespace pod-network-test-627 02/25/23 11:42:36.769 + STEP: creating a selector 02/25/23 11:42:36.769 + STEP: Creating the service pods in kubernetes 02/25/23 11:42:36.77 + Feb 25 11:42:36.770: INFO: Waiting up to 10m0s for all (but 0) nodes to be schedulable + Feb 25 11:42:36.822: INFO: Waiting up to 5m0s for pod "netserver-0" in namespace "pod-network-test-627" to be "running and ready" + Feb 25 11:42:36.828: INFO: Pod "netserver-0": Phase="Pending", Reason="", readiness=false. Elapsed: 6.411857ms + Feb 25 11:42:36.828: INFO: The phase of Pod netserver-0 is Pending, waiting for it to be Running (with Ready = true) + Feb 25 11:42:38.834: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 2.012230024s + Feb 25 11:42:38.834: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:42:40.895: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 4.073195665s + Feb 25 11:42:40.895: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:42:42.835: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 6.01283904s + Feb 25 11:42:42.835: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:42:44.834: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 8.012754701s + Feb 25 11:42:44.835: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:42:46.836: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 10.014398809s + Feb 25 11:42:46.836: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:42:48.835: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 12.013032378s + Feb 25 11:42:48.835: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:42:50.836: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 14.014072923s + Feb 25 11:42:50.836: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:42:52.837: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 16.015405894s + Feb 25 11:42:52.837: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:42:54.837: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 18.015591357s + Feb 25 11:42:54.837: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:42:56.838: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=false. Elapsed: 20.015848659s + Feb 25 11:42:56.838: INFO: The phase of Pod netserver-0 is Running (Ready = false) + Feb 25 11:42:58.835: INFO: Pod "netserver-0": Phase="Running", Reason="", readiness=true. Elapsed: 22.013461164s + Feb 25 11:42:58.835: INFO: The phase of Pod netserver-0 is Running (Ready = true) + Feb 25 11:42:58.835: INFO: Pod "netserver-0" satisfied condition "running and ready" + Feb 25 11:42:58.840: INFO: Waiting up to 5m0s for pod "netserver-1" in namespace "pod-network-test-627" to be "running and ready" + Feb 25 11:42:58.846: INFO: Pod "netserver-1": Phase="Running", Reason="", readiness=true. Elapsed: 5.390314ms + Feb 25 11:42:58.847: INFO: The phase of Pod netserver-1 is Running (Ready = true) + Feb 25 11:42:58.848: INFO: Pod "netserver-1" satisfied condition "running and ready" + Feb 25 11:42:58.854: INFO: Waiting up to 5m0s for pod "netserver-2" in namespace "pod-network-test-627" to be "running and ready" + Feb 25 11:42:58.861: INFO: Pod "netserver-2": Phase="Running", Reason="", readiness=true. Elapsed: 6.437109ms + Feb 25 11:42:58.861: INFO: The phase of Pod netserver-2 is Running (Ready = true) + Feb 25 11:42:58.861: INFO: Pod "netserver-2" satisfied condition "running and ready" + STEP: Creating test pods 02/25/23 11:42:58.866 + Feb 25 11:42:58.895: INFO: Waiting up to 5m0s for pod "test-container-pod" in namespace "pod-network-test-627" to be "running" + Feb 25 11:42:58.917: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 22.657371ms + Feb 25 11:43:00.930: INFO: Pod "test-container-pod": Phase="Pending", Reason="", readiness=false. Elapsed: 2.034823785s + Feb 25 11:43:02.929: INFO: Pod "test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 4.034316993s + Feb 25 11:43:02.929: INFO: Pod "test-container-pod" satisfied condition "running" + Feb 25 11:43:02.934: INFO: Waiting up to 5m0s for pod "host-test-container-pod" in namespace "pod-network-test-627" to be "running" + Feb 25 11:43:02.956: INFO: Pod "host-test-container-pod": Phase="Running", Reason="", readiness=true. Elapsed: 21.329021ms + Feb 25 11:43:02.956: INFO: Pod "host-test-container-pod" satisfied condition "running" + Feb 25 11:43:02.962: INFO: Setting MaxTries for pod polling to 39 for networking test based on endpoint count 3 + Feb 25 11:43:02.962: INFO: Going to poll 10.233.65.137 on port 8083 at least 0 times, with a maximum of 39 tries before failing + Feb 25 11:43:02.975: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.233.65.137:8083/hostName | grep -v '^\s*$'] Namespace:pod-network-test-627 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 11:43:02.975: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 11:43:02.976: INFO: ExecWithOptions: Clientset creation + Feb 25 11:43:02.977: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-627/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+--max-time+15+--connect-timeout+1+http%3A%2F%2F10.233.65.137%3A8083%2FhostName+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) + Feb 25 11:43:03.184: INFO: Found all 1 expected endpoints: [netserver-0] + Feb 25 11:43:03.184: INFO: Going to poll 10.233.64.174 on port 8083 at least 0 times, with a maximum of 39 tries before failing + Feb 25 11:43:03.191: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.233.64.174:8083/hostName | grep -v '^\s*$'] Namespace:pod-network-test-627 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 11:43:03.191: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 11:43:03.192: INFO: ExecWithOptions: Clientset creation + Feb 25 11:43:03.192: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-627/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+--max-time+15+--connect-timeout+1+http%3A%2F%2F10.233.64.174%3A8083%2FhostName+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) + Feb 25 11:43:03.328: INFO: Found all 1 expected endpoints: [netserver-1] + Feb 25 11:43:03.328: INFO: Going to poll 10.233.66.35 on port 8083 at least 0 times, with a maximum of 39 tries before failing + Feb 25 11:43:03.333: INFO: ExecWithOptions {Command:[/bin/sh -c curl -g -q -s --max-time 15 --connect-timeout 1 http://10.233.66.35:8083/hostName | grep -v '^\s*$'] Namespace:pod-network-test-627 PodName:host-test-container-pod ContainerName:agnhost-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} + Feb 25 11:43:03.334: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + Feb 25 11:43:03.335: INFO: ExecWithOptions: Clientset creation + Feb 25 11:43:03.335: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/pod-network-test-627/pods/host-test-container-pod/exec?command=%2Fbin%2Fsh&command=-c&command=curl+-g+-q+-s+--max-time+15+--connect-timeout+1+http%3A%2F%2F10.233.66.35%3A8083%2FhostName+%7C+grep+-v+%27%5E%5Cs%2A%24%27&container=agnhost-container&container=agnhost-container&stderr=true&stdout=true) + Feb 25 11:43:03.417: INFO: Found all 1 expected endpoints: [netserver-2] + [AfterEach] [sig-network] Networking test/e2e/framework/framework.go:187 - Dec 14 17:10:46.239: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "endpointslice-5410" for this suite. 12/14/22 17:10:46.245 + Feb 25 11:43:03.417: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "pod-network-test-627" for this suite. 02/25/23 11:43:03.427 << End Captured GinkgoWriter Output ------------------------------ -SSSSS +SSS ------------------------------ -[sig-storage] EmptyDir volumes - pod should support shared volumes between containers [Conformance] - test/e2e/common/storage/empty_dir.go:226 -[BeforeEach] [sig-storage] EmptyDir volumes +[sig-api-machinery] ResourceQuota + should verify ResourceQuota with best effort scope. [Conformance] + test/e2e/apimachinery/resource_quota.go:793 +[BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:10:46.253 -Dec 14 17:10:46.253: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename emptydir 12/14/22 17:10:46.257 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:10:46.277 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:10:46.28 -[It] pod should support shared volumes between containers [Conformance] - test/e2e/common/storage/empty_dir.go:226 -STEP: Creating Pod 12/14/22 17:10:46.284 -Dec 14 17:10:46.296: INFO: Waiting up to 5m0s for pod "pod-sharedvolume-b0f86afb-6885-4ed9-a8e4-0fe760e28107" in namespace "emptydir-1278" to be "running" -Dec 14 17:10:46.300: INFO: Pod "pod-sharedvolume-b0f86afb-6885-4ed9-a8e4-0fe760e28107": Phase="Pending", Reason="", readiness=false. Elapsed: 4.135288ms -Dec 14 17:10:48.308: INFO: Pod "pod-sharedvolume-b0f86afb-6885-4ed9-a8e4-0fe760e28107": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011248765s -Dec 14 17:10:50.308: INFO: Pod "pod-sharedvolume-b0f86afb-6885-4ed9-a8e4-0fe760e28107": Phase="Running", Reason="", readiness=false. Elapsed: 4.011621729s -Dec 14 17:10:50.308: INFO: Pod "pod-sharedvolume-b0f86afb-6885-4ed9-a8e4-0fe760e28107" satisfied condition "running" -STEP: Reading file content from the nginx-container 12/14/22 17:10:50.308 -Dec 14 17:10:50.308: INFO: ExecWithOptions {Command:[/bin/sh -c cat /usr/share/volumeshare/shareddata.txt] Namespace:emptydir-1278 PodName:pod-sharedvolume-b0f86afb-6885-4ed9-a8e4-0fe760e28107 ContainerName:busybox-main-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} -Dec 14 17:10:50.309: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -Dec 14 17:10:50.310: INFO: ExecWithOptions: Clientset creation -Dec 14 17:10:50.310: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/emptydir-1278/pods/pod-sharedvolume-b0f86afb-6885-4ed9-a8e4-0fe760e28107/exec?command=%2Fbin%2Fsh&command=-c&command=cat+%2Fusr%2Fshare%2Fvolumeshare%2Fshareddata.txt&container=busybox-main-container&container=busybox-main-container&stderr=true&stdout=true) -Dec 14 17:10:50.443: INFO: Exec stderr: "" -[AfterEach] [sig-storage] EmptyDir volumes +STEP: Creating a kubernetes client 02/25/23 11:43:03.438 +Feb 25 11:43:03.438: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename resourcequota 02/25/23 11:43:03.44 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:43:03.49 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:43:03.496 +[It] should verify ResourceQuota with best effort scope. [Conformance] + test/e2e/apimachinery/resource_quota.go:793 +STEP: Creating a ResourceQuota with best effort scope 02/25/23 11:43:03.501 +STEP: Ensuring ResourceQuota status is calculated 02/25/23 11:43:03.512 +STEP: Creating a ResourceQuota with not best effort scope 02/25/23 11:43:05.522 +STEP: Ensuring ResourceQuota status is calculated 02/25/23 11:43:05.53 +STEP: Creating a best-effort pod 02/25/23 11:43:07.537 +STEP: Ensuring resource quota with best effort scope captures the pod usage 02/25/23 11:43:07.563 +STEP: Ensuring resource quota with not best effort ignored the pod usage 02/25/23 11:43:09.571 +STEP: Deleting the pod 02/25/23 11:43:11.579 +STEP: Ensuring resource quota status released the pod usage 02/25/23 11:43:11.615 +STEP: Creating a not best-effort pod 02/25/23 11:43:13.64 +STEP: Ensuring resource quota with not best effort scope captures the pod usage 02/25/23 11:43:13.66 +STEP: Ensuring resource quota with best effort scope ignored the pod usage 02/25/23 11:43:15.668 +STEP: Deleting the pod 02/25/23 11:43:17.676 +STEP: Ensuring resource quota status released the pod usage 02/25/23 11:43:17.698 +[AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 -Dec 14 17:10:50.443: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "emptydir-1278" for this suite. 12/14/22 17:10:50.45 -{"msg":"PASSED [sig-storage] EmptyDir volumes pod should support shared volumes between containers [Conformance]","completed":358,"skipped":6646,"failed":0} +Feb 25 11:43:19.707: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "resourcequota-5848" for this suite. 02/25/23 11:43:19.714 +{"msg":"PASSED [sig-api-machinery] ResourceQuota should verify ResourceQuota with best effort scope. [Conformance]","completed":358,"skipped":6655,"failed":0} ------------------------------ -• [4.207 seconds] -[sig-storage] EmptyDir volumes -test/e2e/common/storage/framework.go:23 - pod should support shared volumes between containers [Conformance] - test/e2e/common/storage/empty_dir.go:226 +• [SLOW TEST] [16.288 seconds] +[sig-api-machinery] ResourceQuota +test/e2e/apimachinery/framework.go:23 + should verify ResourceQuota with best effort scope. [Conformance] + test/e2e/apimachinery/resource_quota.go:793 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-storage] EmptyDir volumes + [BeforeEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:10:46.253 - Dec 14 17:10:46.253: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename emptydir 12/14/22 17:10:46.257 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:10:46.277 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:10:46.28 - [It] pod should support shared volumes between containers [Conformance] - test/e2e/common/storage/empty_dir.go:226 - STEP: Creating Pod 12/14/22 17:10:46.284 - Dec 14 17:10:46.296: INFO: Waiting up to 5m0s for pod "pod-sharedvolume-b0f86afb-6885-4ed9-a8e4-0fe760e28107" in namespace "emptydir-1278" to be "running" - Dec 14 17:10:46.300: INFO: Pod "pod-sharedvolume-b0f86afb-6885-4ed9-a8e4-0fe760e28107": Phase="Pending", Reason="", readiness=false. Elapsed: 4.135288ms - Dec 14 17:10:48.308: INFO: Pod "pod-sharedvolume-b0f86afb-6885-4ed9-a8e4-0fe760e28107": Phase="Pending", Reason="", readiness=false. Elapsed: 2.011248765s - Dec 14 17:10:50.308: INFO: Pod "pod-sharedvolume-b0f86afb-6885-4ed9-a8e4-0fe760e28107": Phase="Running", Reason="", readiness=false. Elapsed: 4.011621729s - Dec 14 17:10:50.308: INFO: Pod "pod-sharedvolume-b0f86afb-6885-4ed9-a8e4-0fe760e28107" satisfied condition "running" - STEP: Reading file content from the nginx-container 12/14/22 17:10:50.308 - Dec 14 17:10:50.308: INFO: ExecWithOptions {Command:[/bin/sh -c cat /usr/share/volumeshare/shareddata.txt] Namespace:emptydir-1278 PodName:pod-sharedvolume-b0f86afb-6885-4ed9-a8e4-0fe760e28107 ContainerName:busybox-main-container Stdin: CaptureStdout:true CaptureStderr:true PreserveWhitespace:false Quiet:false} - Dec 14 17:10:50.309: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - Dec 14 17:10:50.310: INFO: ExecWithOptions: Clientset creation - Dec 14 17:10:50.310: INFO: ExecWithOptions: execute(POST https://10.233.0.1:443/api/v1/namespaces/emptydir-1278/pods/pod-sharedvolume-b0f86afb-6885-4ed9-a8e4-0fe760e28107/exec?command=%2Fbin%2Fsh&command=-c&command=cat+%2Fusr%2Fshare%2Fvolumeshare%2Fshareddata.txt&container=busybox-main-container&container=busybox-main-container&stderr=true&stdout=true) - Dec 14 17:10:50.443: INFO: Exec stderr: "" - [AfterEach] [sig-storage] EmptyDir volumes + STEP: Creating a kubernetes client 02/25/23 11:43:03.438 + Feb 25 11:43:03.438: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename resourcequota 02/25/23 11:43:03.44 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:43:03.49 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:43:03.496 + [It] should verify ResourceQuota with best effort scope. [Conformance] + test/e2e/apimachinery/resource_quota.go:793 + STEP: Creating a ResourceQuota with best effort scope 02/25/23 11:43:03.501 + STEP: Ensuring ResourceQuota status is calculated 02/25/23 11:43:03.512 + STEP: Creating a ResourceQuota with not best effort scope 02/25/23 11:43:05.522 + STEP: Ensuring ResourceQuota status is calculated 02/25/23 11:43:05.53 + STEP: Creating a best-effort pod 02/25/23 11:43:07.537 + STEP: Ensuring resource quota with best effort scope captures the pod usage 02/25/23 11:43:07.563 + STEP: Ensuring resource quota with not best effort ignored the pod usage 02/25/23 11:43:09.571 + STEP: Deleting the pod 02/25/23 11:43:11.579 + STEP: Ensuring resource quota status released the pod usage 02/25/23 11:43:11.615 + STEP: Creating a not best-effort pod 02/25/23 11:43:13.64 + STEP: Ensuring resource quota with not best effort scope captures the pod usage 02/25/23 11:43:13.66 + STEP: Ensuring resource quota with best effort scope ignored the pod usage 02/25/23 11:43:15.668 + STEP: Deleting the pod 02/25/23 11:43:17.676 + STEP: Ensuring resource quota status released the pod usage 02/25/23 11:43:17.698 + [AfterEach] [sig-api-machinery] ResourceQuota test/e2e/framework/framework.go:187 - Dec 14 17:10:50.443: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "emptydir-1278" for this suite. 12/14/22 17:10:50.45 + Feb 25 11:43:19.707: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "resourcequota-5848" for this suite. 02/25/23 11:43:19.714 << End Captured GinkgoWriter Output ------------------------------ -SSS +SSSSSSSSSSSSSSSSSSS ------------------------------ -[sig-apps] CronJob - should not schedule new jobs when ForbidConcurrent [Slow] [Conformance] - test/e2e/apps/cronjob.go:124 -[BeforeEach] [sig-apps] CronJob +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + should be able to deny custom resource creation, update and deletion [Conformance] + test/e2e/apimachinery/webhook.go:220 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:10:50.463 -Dec 14 17:10:50.463: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename cronjob 12/14/22 17:10:50.466 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:10:50.491 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:10:50.497 -[It] should not schedule new jobs when ForbidConcurrent [Slow] [Conformance] - test/e2e/apps/cronjob.go:124 -STEP: Creating a ForbidConcurrent cronjob 12/14/22 17:10:50.501 -STEP: Ensuring a job is scheduled 12/14/22 17:10:50.511 -STEP: Ensuring exactly one is scheduled 12/14/22 17:11:00.518 -STEP: Ensuring exactly one running job exists by listing jobs explicitly 12/14/22 17:11:00.523 -STEP: Ensuring no more jobs are scheduled 12/14/22 17:11:00.529 -STEP: Removing cronjob 12/14/22 17:16:00.545 -[AfterEach] [sig-apps] CronJob +STEP: Creating a kubernetes client 02/25/23 11:43:19.733 +Feb 25 11:43:19.734: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename webhook 02/25/23 11:43:19.736 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:43:19.77 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:43:19.774 +[BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 +STEP: Setting up server cert 02/25/23 11:43:19.797 +STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 11:43:21.367 +STEP: Deploying the webhook pod 02/25/23 11:43:21.41 +STEP: Wait for the deployment to be ready 02/25/23 11:43:21.461 +Feb 25 11:43:21.489: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set +Feb 25 11:43:23.541: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 11, 43, 21, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 43, 21, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 11, 43, 21, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 43, 21, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} +STEP: Deploying the webhook service 02/25/23 11:43:25.548 +STEP: Verifying the service has paired with the endpoint 02/25/23 11:43:25.579 +Feb 25 11:43:26.579: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 +[It] should be able to deny custom resource creation, update and deletion [Conformance] + test/e2e/apimachinery/webhook.go:220 +Feb 25 11:43:26.585: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Registering the custom resource webhook via the AdmissionRegistration API 02/25/23 11:43:27.099 +STEP: Creating a custom resource that should be denied by the webhook 02/25/23 11:43:27.136 +STEP: Creating a custom resource whose deletion would be denied by the webhook 02/25/23 11:43:29.298 +STEP: Updating the custom resource with disallowed data should be denied 02/25/23 11:43:29.314 +STEP: Deleting the custom resource should be denied 02/25/23 11:43:29.331 +STEP: Remove the offending key and value from the custom resource data 02/25/23 11:43:29.343 +STEP: Deleting the updated custom resource should be successful 02/25/23 11:43:29.374 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 -Dec 14 17:16:00.557: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "cronjob-2907" for this suite. 12/14/22 17:16:00.571 -{"msg":"PASSED [sig-apps] CronJob should not schedule new jobs when ForbidConcurrent [Slow] [Conformance]","completed":359,"skipped":6649,"failed":0} +Feb 25 11:43:29.918: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "webhook-1999" for this suite. 02/25/23 11:43:29.927 +STEP: Destroying namespace "webhook-1999-markers" for this suite. 02/25/23 11:43:29.938 +[AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 +{"msg":"PASSED [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] should be able to deny custom resource creation, update and deletion [Conformance]","completed":359,"skipped":6674,"failed":0} ------------------------------ -• [SLOW TEST] [310.118 seconds] -[sig-apps] CronJob -test/e2e/apps/framework.go:23 - should not schedule new jobs when ForbidConcurrent [Slow] [Conformance] - test/e2e/apps/cronjob.go:124 +• [SLOW TEST] [10.279 seconds] +[sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] +test/e2e/apimachinery/framework.go:23 + should be able to deny custom resource creation, update and deletion [Conformance] + test/e2e/apimachinery/webhook.go:220 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-apps] CronJob + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:10:50.463 - Dec 14 17:10:50.463: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename cronjob 12/14/22 17:10:50.466 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:10:50.491 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:10:50.497 - [It] should not schedule new jobs when ForbidConcurrent [Slow] [Conformance] - test/e2e/apps/cronjob.go:124 - STEP: Creating a ForbidConcurrent cronjob 12/14/22 17:10:50.501 - STEP: Ensuring a job is scheduled 12/14/22 17:10:50.511 - STEP: Ensuring exactly one is scheduled 12/14/22 17:11:00.518 - STEP: Ensuring exactly one running job exists by listing jobs explicitly 12/14/22 17:11:00.523 - STEP: Ensuring no more jobs are scheduled 12/14/22 17:11:00.529 - STEP: Removing cronjob 12/14/22 17:16:00.545 - [AfterEach] [sig-apps] CronJob + STEP: Creating a kubernetes client 02/25/23 11:43:19.733 + Feb 25 11:43:19.734: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename webhook 02/25/23 11:43:19.736 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:43:19.77 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:43:19.774 + [BeforeEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:89 + STEP: Setting up server cert 02/25/23 11:43:19.797 + STEP: Create role binding to let webhook read extension-apiserver-authentication 02/25/23 11:43:21.367 + STEP: Deploying the webhook pod 02/25/23 11:43:21.41 + STEP: Wait for the deployment to be ready 02/25/23 11:43:21.461 + Feb 25 11:43:21.489: INFO: deployment "sample-webhook-deployment" doesn't have the required revision set + Feb 25 11:43:23.541: INFO: deployment status: v1.DeploymentStatus{ObservedGeneration:1, Replicas:1, UpdatedReplicas:1, ReadyReplicas:0, AvailableReplicas:0, UnavailableReplicas:1, Conditions:[]v1.DeploymentCondition{v1.DeploymentCondition{Type:"Available", Status:"False", LastUpdateTime:time.Date(2023, time.February, 25, 11, 43, 21, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 43, 21, 0, time.Local), Reason:"MinimumReplicasUnavailable", Message:"Deployment does not have minimum availability."}, v1.DeploymentCondition{Type:"Progressing", Status:"True", LastUpdateTime:time.Date(2023, time.February, 25, 11, 43, 21, 0, time.Local), LastTransitionTime:time.Date(2023, time.February, 25, 11, 43, 21, 0, time.Local), Reason:"ReplicaSetUpdated", Message:"ReplicaSet \"sample-webhook-deployment-5d85dd8cdb\" is progressing."}}, CollisionCount:(*int32)(nil)} + STEP: Deploying the webhook service 02/25/23 11:43:25.548 + STEP: Verifying the service has paired with the endpoint 02/25/23 11:43:25.579 + Feb 25 11:43:26.579: INFO: Waiting for amount of service:e2e-test-webhook endpoints to be 1 + [It] should be able to deny custom resource creation, update and deletion [Conformance] + test/e2e/apimachinery/webhook.go:220 + Feb 25 11:43:26.585: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Registering the custom resource webhook via the AdmissionRegistration API 02/25/23 11:43:27.099 + STEP: Creating a custom resource that should be denied by the webhook 02/25/23 11:43:27.136 + STEP: Creating a custom resource whose deletion would be denied by the webhook 02/25/23 11:43:29.298 + STEP: Updating the custom resource with disallowed data should be denied 02/25/23 11:43:29.314 + STEP: Deleting the custom resource should be denied 02/25/23 11:43:29.331 + STEP: Remove the offending key and value from the custom resource data 02/25/23 11:43:29.343 + STEP: Deleting the updated custom resource should be successful 02/25/23 11:43:29.374 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] test/e2e/framework/framework.go:187 - Dec 14 17:16:00.557: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "cronjob-2907" for this suite. 12/14/22 17:16:00.571 + Feb 25 11:43:29.918: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "webhook-1999" for this suite. 02/25/23 11:43:29.927 + STEP: Destroying namespace "webhook-1999-markers" for this suite. 02/25/23 11:43:29.938 + [AfterEach] [sig-api-machinery] AdmissionWebhook [Privileged:ClusterAdmin] + test/e2e/apimachinery/webhook.go:104 << End Captured GinkgoWriter Output ------------------------------ -[sig-network] Service endpoints latency - should not be very high [Conformance] - test/e2e/network/service_latency.go:59 -[BeforeEach] [sig-network] Service endpoints latency +SSS +------------------------------ +[sig-api-machinery] Namespaces [Serial] + should ensure that all services are removed when a namespace is deleted [Conformance] + test/e2e/apimachinery/namespace.go:250 +[BeforeEach] [sig-api-machinery] Namespaces [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:16:00.581 -Dec 14 17:16:00.582: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename svc-latency 12/14/22 17:16:00.588 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:16:00.65 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:16:00.654 -[It] should not be very high [Conformance] - test/e2e/network/service_latency.go:59 -Dec 14 17:16:00.657: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: creating replication controller svc-latency-rc in namespace svc-latency-4007 12/14/22 17:16:00.66 -I1214 17:16:00.670883 14 runners.go:193] Created replication controller with name: svc-latency-rc, namespace: svc-latency-4007, replica count: 1 -I1214 17:16:01.722077 14 runners.go:193] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -I1214 17:16:02.722385 14 runners.go:193] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady -Dec 14 17:16:02.837: INFO: Created: latency-svc-qwk5l -Dec 14 17:16:02.874: INFO: Got endpoints: latency-svc-qwk5l [51.950228ms] -Dec 14 17:16:02.895: INFO: Created: latency-svc-xns98 -Dec 14 17:16:02.906: INFO: Got endpoints: latency-svc-xns98 [31.396245ms] -Dec 14 17:16:03.027: INFO: Created: latency-svc-kk4xc -Dec 14 17:16:03.028: INFO: Created: latency-svc-vq4vn -Dec 14 17:16:03.028: INFO: Created: latency-svc-hqgx4 -Dec 14 17:16:03.028: INFO: Created: latency-svc-lp6rw -Dec 14 17:16:03.029: INFO: Created: latency-svc-8lzd5 -Dec 14 17:16:03.030: INFO: Created: latency-svc-7dc67 -Dec 14 17:16:03.035: INFO: Created: latency-svc-8t5c9 -Dec 14 17:16:03.036: INFO: Created: latency-svc-nzcxn -Dec 14 17:16:03.036: INFO: Created: latency-svc-vrhnk -Dec 14 17:16:03.036: INFO: Created: latency-svc-92vlc -Dec 14 17:16:03.041: INFO: Got endpoints: latency-svc-7dc67 [165.788645ms] -Dec 14 17:16:03.055: INFO: Created: latency-svc-thjr2 -Dec 14 17:16:03.056: INFO: Created: latency-svc-wrmr9 -Dec 14 17:16:03.063: INFO: Created: latency-svc-kjkrj -Dec 14 17:16:03.133: INFO: Got endpoints: latency-svc-thjr2 [257.992284ms] -Dec 14 17:16:03.137: INFO: Got endpoints: latency-svc-wrmr9 [262.079176ms] -Dec 14 17:16:03.142: INFO: Created: latency-svc-7b7lt -Dec 14 17:16:03.143: INFO: Created: latency-svc-4pntf -Dec 14 17:16:03.146: INFO: Got endpoints: latency-svc-8lzd5 [240.271249ms] -Dec 14 17:16:03.157: INFO: Got endpoints: latency-svc-kk4xc [281.270241ms] -Dec 14 17:16:03.163: INFO: Got endpoints: latency-svc-lp6rw [287.891065ms] -Dec 14 17:16:03.167: INFO: Created: latency-svc-l8d8b -Dec 14 17:16:03.186: INFO: Got endpoints: latency-svc-hqgx4 [309.272446ms] -Dec 14 17:16:03.193: INFO: Got endpoints: latency-svc-vq4vn [317.421903ms] -Dec 14 17:16:03.194: INFO: Got endpoints: latency-svc-92vlc [318.867562ms] -Dec 14 17:16:03.195: INFO: Got endpoints: latency-svc-8t5c9 [319.346279ms] -Dec 14 17:16:03.202: INFO: Created: latency-svc-5ljws -Dec 14 17:16:03.213: INFO: Got endpoints: latency-svc-nzcxn [337.007557ms] -Dec 14 17:16:03.231: INFO: Got endpoints: latency-svc-vrhnk [354.620515ms] -Dec 14 17:16:03.231: INFO: Got endpoints: latency-svc-4pntf [355.575227ms] -Dec 14 17:16:03.238: INFO: Created: latency-svc-4qm9z -Dec 14 17:16:03.239: INFO: Created: latency-svc-dzpwx -Dec 14 17:16:03.259: INFO: Got endpoints: latency-svc-7b7lt [382.685597ms] -Dec 14 17:16:03.270: INFO: Got endpoints: latency-svc-l8d8b [394.641869ms] -Dec 14 17:16:03.274: INFO: Got endpoints: latency-svc-5ljws [141.208819ms] -Dec 14 17:16:03.275: INFO: Got endpoints: latency-svc-kjkrj [234.213126ms] -Dec 14 17:16:03.279: INFO: Got endpoints: latency-svc-4qm9z [141.994896ms] -Dec 14 17:16:03.288: INFO: Created: latency-svc-p58cv -Dec 14 17:16:03.309: INFO: Got endpoints: latency-svc-dzpwx [163.230592ms] -Dec 14 17:16:03.341: INFO: Got endpoints: latency-svc-p58cv [183.797748ms] -Dec 14 17:16:03.354: INFO: Created: latency-svc-4hg2p -Dec 14 17:16:03.355: INFO: Created: latency-svc-2rrjd -Dec 14 17:16:03.360: INFO: Got endpoints: latency-svc-2rrjd [196.846151ms] -Dec 14 17:16:03.368: INFO: Got endpoints: latency-svc-4hg2p [182.137895ms] -Dec 14 17:16:03.385: INFO: Created: latency-svc-8rwz9 -Dec 14 17:16:03.398: INFO: Got endpoints: latency-svc-8rwz9 [203.596618ms] -Dec 14 17:16:03.408: INFO: Created: latency-svc-jrjlh -Dec 14 17:16:03.415: INFO: Got endpoints: latency-svc-jrjlh [221.701232ms] -Dec 14 17:16:03.418: INFO: Created: latency-svc-tkscz -Dec 14 17:16:03.434: INFO: Created: latency-svc-mws6h -Dec 14 17:16:03.442: INFO: Got endpoints: latency-svc-tkscz [246.711469ms] -Dec 14 17:16:03.448: INFO: Created: latency-svc-c8hw7 -Dec 14 17:16:03.461: INFO: Created: latency-svc-2rkf8 -Dec 14 17:16:03.474: INFO: Got endpoints: latency-svc-mws6h [261.528179ms] -Dec 14 17:16:03.488: INFO: Got endpoints: latency-svc-c8hw7 [255.97453ms] -Dec 14 17:16:03.494: INFO: Created: latency-svc-txpz8 -Dec 14 17:16:03.515: INFO: Got endpoints: latency-svc-2rkf8 [283.741899ms] -Dec 14 17:16:03.524: INFO: Got endpoints: latency-svc-txpz8 [253.527213ms] -Dec 14 17:16:03.557: INFO: Created: latency-svc-c98qw -Dec 14 17:16:03.574: INFO: Got endpoints: latency-svc-c98qw [299.208111ms] -Dec 14 17:16:03.577: INFO: Created: latency-svc-bvbqk -Dec 14 17:16:03.604: INFO: Got endpoints: latency-svc-bvbqk [329.47754ms] -Dec 14 17:16:03.620: INFO: Created: latency-svc-wnvrr -Dec 14 17:16:03.645: INFO: Got endpoints: latency-svc-wnvrr [365.509196ms] -Dec 14 17:16:03.652: INFO: Created: latency-svc-5n8rf -Dec 14 17:16:03.665: INFO: Created: latency-svc-nxbr6 -Dec 14 17:16:03.672: INFO: Got endpoints: latency-svc-5n8rf [362.204865ms] -Dec 14 17:16:03.692: INFO: Got endpoints: latency-svc-nxbr6 [350.53158ms] -Dec 14 17:16:03.699: INFO: Created: latency-svc-ptfwd -Dec 14 17:16:03.705: INFO: Got endpoints: latency-svc-ptfwd [345.118266ms] -Dec 14 17:16:03.709: INFO: Created: latency-svc-78ksf -Dec 14 17:16:03.722: INFO: Created: latency-svc-7k7ds -Dec 14 17:16:03.722: INFO: Got endpoints: latency-svc-78ksf [353.648852ms] -Dec 14 17:16:03.730: INFO: Created: latency-svc-mczqt -Dec 14 17:16:03.730: INFO: Got endpoints: latency-svc-7k7ds [332.274215ms] -Dec 14 17:16:03.742: INFO: Created: latency-svc-bkzp6 -Dec 14 17:16:03.748: INFO: Created: latency-svc-hdcr6 -Dec 14 17:16:03.751: INFO: Created: latency-svc-76vtw -Dec 14 17:16:03.757: INFO: Got endpoints: latency-svc-mczqt [341.811913ms] -Dec 14 17:16:03.783: INFO: Created: latency-svc-zmwmw -Dec 14 17:16:03.784: INFO: Got endpoints: latency-svc-hdcr6 [342.52653ms] -Dec 14 17:16:03.785: INFO: Got endpoints: latency-svc-76vtw [310.229673ms] -Dec 14 17:16:03.785: INFO: Got endpoints: latency-svc-bkzp6 [525.76334ms] -Dec 14 17:16:03.794: INFO: Created: latency-svc-4sjvg -Dec 14 17:16:03.823: INFO: Created: latency-svc-vcnph -Dec 14 17:16:03.824: INFO: Got endpoints: latency-svc-zmwmw [336.006585ms] -Dec 14 17:16:03.824: INFO: Got endpoints: latency-svc-4sjvg [309.409449ms] -Dec 14 17:16:03.860: INFO: Got endpoints: latency-svc-vcnph [336.170935ms] -Dec 14 17:16:03.868: INFO: Created: latency-svc-8ztnf -Dec 14 17:16:03.877: INFO: Created: latency-svc-nllnl -Dec 14 17:16:03.878: INFO: Got endpoints: latency-svc-8ztnf [304.480559ms] -Dec 14 17:16:03.889: INFO: Got endpoints: latency-svc-nllnl [284.665018ms] -Dec 14 17:16:03.900: INFO: Created: latency-svc-7m8gx -Dec 14 17:16:03.927: INFO: Got endpoints: latency-svc-7m8gx [281.859048ms] -Dec 14 17:16:03.940: INFO: Created: latency-svc-kx4x5 -Dec 14 17:16:03.940: INFO: Created: latency-svc-5trjp -Dec 14 17:16:03.946: INFO: Got endpoints: latency-svc-kx4x5 [273.54668ms] -Dec 14 17:16:03.969: INFO: Created: latency-svc-9ms4v -Dec 14 17:16:03.990: INFO: Got endpoints: latency-svc-5trjp [297.766123ms] -Dec 14 17:16:03.996: INFO: Created: latency-svc-c2mxz -Dec 14 17:16:04.004: INFO: Created: latency-svc-qxbsp -Dec 14 17:16:04.030: INFO: Created: latency-svc-cblft -Dec 14 17:16:04.042: INFO: Got endpoints: latency-svc-9ms4v [336.421299ms] -Dec 14 17:16:04.055: INFO: Created: latency-svc-r24vc -Dec 14 17:16:04.065: INFO: Created: latency-svc-2kdkx -Dec 14 17:16:04.076: INFO: Created: latency-svc-p7hwp -Dec 14 17:16:04.099: INFO: Got endpoints: latency-svc-qxbsp [376.852681ms] -Dec 14 17:16:04.112: INFO: Created: latency-svc-rf99f -Dec 14 17:16:04.130: INFO: Created: latency-svc-dmt2n -Dec 14 17:16:04.148: INFO: Got endpoints: latency-svc-c2mxz [417.611771ms] -Dec 14 17:16:04.149: INFO: Created: latency-svc-6qpnz -Dec 14 17:16:04.164: INFO: Created: latency-svc-z8vn5 -Dec 14 17:16:04.169: INFO: Created: latency-svc-5pmw7 -Dec 14 17:16:04.212: INFO: Created: latency-svc-k7cng -Dec 14 17:16:04.238: INFO: Got endpoints: latency-svc-cblft [480.530728ms] -Dec 14 17:16:04.249: INFO: Got endpoints: latency-svc-2kdkx [464.683146ms] -Dec 14 17:16:04.272: INFO: Created: latency-svc-xc5sn -Dec 14 17:16:04.276: INFO: Created: latency-svc-ls99t -Dec 14 17:16:04.293: INFO: Got endpoints: latency-svc-r24vc [508.218847ms] -Dec 14 17:16:04.365: INFO: Got endpoints: latency-svc-p7hwp [580.957445ms] -Dec 14 17:16:04.396: INFO: Created: latency-svc-r5frc -Dec 14 17:16:04.398: INFO: Got endpoints: latency-svc-rf99f [574.349617ms] -Dec 14 17:16:04.403: INFO: Created: latency-svc-w8jk6 -Dec 14 17:16:04.412: INFO: Created: latency-svc-xhd8p -Dec 14 17:16:04.442: INFO: Got endpoints: latency-svc-dmt2n [617.396396ms] -Dec 14 17:16:04.459: INFO: Created: latency-svc-c4tbm -Dec 14 17:16:04.474: INFO: Created: latency-svc-mvf9p -Dec 14 17:16:04.480: INFO: Created: latency-svc-mjl22 -Dec 14 17:16:04.504: INFO: Created: latency-svc-nzcsn -Dec 14 17:16:04.514: INFO: Got endpoints: latency-svc-6qpnz [653.692434ms] -Dec 14 17:16:04.525: INFO: Created: latency-svc-w4dkn -Dec 14 17:16:04.533: INFO: Created: latency-svc-kndgt -Dec 14 17:16:04.557: INFO: Got endpoints: latency-svc-z8vn5 [678.87874ms] -Dec 14 17:16:04.566: INFO: Created: latency-svc-5q9qg -Dec 14 17:16:04.580: INFO: Created: latency-svc-vbcnl -Dec 14 17:16:04.596: INFO: Got endpoints: latency-svc-5pmw7 [706.774845ms] -Dec 14 17:16:04.618: INFO: Created: latency-svc-lr569 -Dec 14 17:16:04.643: INFO: Got endpoints: latency-svc-k7cng [715.473646ms] -Dec 14 17:16:04.660: INFO: Created: latency-svc-m2fd9 -Dec 14 17:16:04.695: INFO: Got endpoints: latency-svc-xc5sn [748.536241ms] -Dec 14 17:16:04.723: INFO: Created: latency-svc-gd89s -Dec 14 17:16:04.745: INFO: Got endpoints: latency-svc-ls99t [755.05949ms] -Dec 14 17:16:04.773: INFO: Created: latency-svc-bs2kc -Dec 14 17:16:04.793: INFO: Got endpoints: latency-svc-r5frc [750.960083ms] -Dec 14 17:16:04.812: INFO: Created: latency-svc-n6ml6 -Dec 14 17:16:04.843: INFO: Got endpoints: latency-svc-w8jk6 [744.640515ms] -Dec 14 17:16:04.867: INFO: Created: latency-svc-6bvgc -Dec 14 17:16:04.895: INFO: Got endpoints: latency-svc-xhd8p [747.051974ms] -Dec 14 17:16:04.917: INFO: Created: latency-svc-ndhn7 -Dec 14 17:16:04.942: INFO: Got endpoints: latency-svc-c4tbm [703.945132ms] -Dec 14 17:16:04.959: INFO: Created: latency-svc-bnrr5 -Dec 14 17:16:04.994: INFO: Got endpoints: latency-svc-mvf9p [744.555911ms] -Dec 14 17:16:05.012: INFO: Created: latency-svc-dvh9h -Dec 14 17:16:05.041: INFO: Got endpoints: latency-svc-mjl22 [747.40927ms] -Dec 14 17:16:05.061: INFO: Created: latency-svc-kwcr4 -Dec 14 17:16:05.190: INFO: Got endpoints: latency-svc-nzcsn [823.091477ms] -Dec 14 17:16:05.193: INFO: Got endpoints: latency-svc-w4dkn [793.324979ms] -Dec 14 17:16:05.196: INFO: Got endpoints: latency-svc-kndgt [754.138026ms] -Dec 14 17:16:05.215: INFO: Created: latency-svc-bxkhg -Dec 14 17:16:05.225: INFO: Created: latency-svc-8ldjj -Dec 14 17:16:05.231: INFO: Created: latency-svc-j6hbc -Dec 14 17:16:05.240: INFO: Got endpoints: latency-svc-5q9qg [725.704781ms] -Dec 14 17:16:05.257: INFO: Created: latency-svc-xj724 -Dec 14 17:16:05.292: INFO: Got endpoints: latency-svc-vbcnl [734.362717ms] -Dec 14 17:16:05.312: INFO: Created: latency-svc-fp5b7 -Dec 14 17:16:05.341: INFO: Got endpoints: latency-svc-lr569 [744.678451ms] -Dec 14 17:16:05.355: INFO: Created: latency-svc-r2cgx -Dec 14 17:16:05.394: INFO: Got endpoints: latency-svc-m2fd9 [751.198596ms] -Dec 14 17:16:05.407: INFO: Created: latency-svc-tnkxc -Dec 14 17:16:05.443: INFO: Got endpoints: latency-svc-gd89s [748.388158ms] -Dec 14 17:16:05.458: INFO: Created: latency-svc-9zqhg -Dec 14 17:16:05.491: INFO: Got endpoints: latency-svc-bs2kc [746.199526ms] -Dec 14 17:16:05.505: INFO: Created: latency-svc-nhjvs -Dec 14 17:16:05.548: INFO: Got endpoints: latency-svc-n6ml6 [754.66886ms] -Dec 14 17:16:05.561: INFO: Created: latency-svc-mz2wn -Dec 14 17:16:05.589: INFO: Got endpoints: latency-svc-6bvgc [744.925734ms] -Dec 14 17:16:05.616: INFO: Created: latency-svc-7p5f4 -Dec 14 17:16:05.655: INFO: Got endpoints: latency-svc-ndhn7 [759.466289ms] -Dec 14 17:16:05.675: INFO: Created: latency-svc-tvdm6 -Dec 14 17:16:05.695: INFO: Got endpoints: latency-svc-bnrr5 [753.495156ms] -Dec 14 17:16:05.714: INFO: Created: latency-svc-zhkkb -Dec 14 17:16:05.738: INFO: Got endpoints: latency-svc-dvh9h [743.928118ms] -Dec 14 17:16:05.785: INFO: Created: latency-svc-k8hwb -Dec 14 17:16:05.795: INFO: Got endpoints: latency-svc-kwcr4 [753.535874ms] -Dec 14 17:16:05.817: INFO: Created: latency-svc-6kp27 -Dec 14 17:16:05.874: INFO: Got endpoints: latency-svc-bxkhg [681.836381ms] -Dec 14 17:16:05.891: INFO: Got endpoints: latency-svc-8ldjj [700.445312ms] -Dec 14 17:16:05.962: INFO: Created: latency-svc-mcvgr -Dec 14 17:16:05.963: INFO: Created: latency-svc-vvsqf -Dec 14 17:16:05.964: INFO: Got endpoints: latency-svc-j6hbc [768.221977ms] -Dec 14 17:16:05.977: INFO: Created: latency-svc-kr2sk -Dec 14 17:16:05.996: INFO: Got endpoints: latency-svc-xj724 [756.148074ms] -Dec 14 17:16:06.034: INFO: Created: latency-svc-cpwzn -Dec 14 17:16:06.045: INFO: Got endpoints: latency-svc-fp5b7 [753.171167ms] -Dec 14 17:16:06.077: INFO: Created: latency-svc-9ljxp -Dec 14 17:16:06.122: INFO: Got endpoints: latency-svc-r2cgx [781.110117ms] -Dec 14 17:16:06.137: INFO: Created: latency-svc-ssjjf -Dec 14 17:16:06.153: INFO: Got endpoints: latency-svc-tnkxc [758.569214ms] -Dec 14 17:16:06.184: INFO: Created: latency-svc-xdsqz -Dec 14 17:16:06.197: INFO: Got endpoints: latency-svc-9zqhg [753.905958ms] -Dec 14 17:16:06.211: INFO: Created: latency-svc-fwkr7 -Dec 14 17:16:06.245: INFO: Got endpoints: latency-svc-nhjvs [753.313339ms] -Dec 14 17:16:06.259: INFO: Created: latency-svc-rtrc2 -Dec 14 17:16:06.296: INFO: Got endpoints: latency-svc-mz2wn [748.5896ms] -Dec 14 17:16:06.312: INFO: Created: latency-svc-7cgb2 -Dec 14 17:16:06.341: INFO: Got endpoints: latency-svc-7p5f4 [752.865812ms] -Dec 14 17:16:06.379: INFO: Created: latency-svc-t5plk -Dec 14 17:16:06.401: INFO: Got endpoints: latency-svc-tvdm6 [746.348404ms] -Dec 14 17:16:06.414: INFO: Created: latency-svc-wc9wl -Dec 14 17:16:06.439: INFO: Got endpoints: latency-svc-zhkkb [743.985635ms] -Dec 14 17:16:06.459: INFO: Created: latency-svc-88vpx -Dec 14 17:16:06.496: INFO: Got endpoints: latency-svc-k8hwb [758.251774ms] -Dec 14 17:16:06.511: INFO: Created: latency-svc-s82xq -Dec 14 17:16:06.539: INFO: Got endpoints: latency-svc-6kp27 [744.700531ms] -Dec 14 17:16:06.555: INFO: Created: latency-svc-9j7g9 -Dec 14 17:16:06.590: INFO: Got endpoints: latency-svc-vvsqf [715.563427ms] -Dec 14 17:16:06.607: INFO: Created: latency-svc-5jjmr -Dec 14 17:16:06.641: INFO: Got endpoints: latency-svc-mcvgr [750.253503ms] -Dec 14 17:16:06.660: INFO: Created: latency-svc-8zmzg -Dec 14 17:16:06.690: INFO: Got endpoints: latency-svc-kr2sk [726.069733ms] -Dec 14 17:16:06.703: INFO: Created: latency-svc-jr26z -Dec 14 17:16:06.741: INFO: Got endpoints: latency-svc-cpwzn [744.226084ms] -Dec 14 17:16:06.755: INFO: Created: latency-svc-z6kb2 -Dec 14 17:16:06.788: INFO: Got endpoints: latency-svc-9ljxp [742.61698ms] -Dec 14 17:16:06.801: INFO: Created: latency-svc-cmnsh -Dec 14 17:16:06.840: INFO: Got endpoints: latency-svc-ssjjf [717.573346ms] -Dec 14 17:16:06.853: INFO: Created: latency-svc-56mq6 -Dec 14 17:16:06.898: INFO: Got endpoints: latency-svc-xdsqz [745.64416ms] -Dec 14 17:16:06.915: INFO: Created: latency-svc-2xvvr -Dec 14 17:16:06.940: INFO: Got endpoints: latency-svc-fwkr7 [743.416321ms] -Dec 14 17:16:06.957: INFO: Created: latency-svc-pks7q -Dec 14 17:16:06.990: INFO: Got endpoints: latency-svc-rtrc2 [745.474333ms] -Dec 14 17:16:07.001: INFO: Created: latency-svc-xfj2x -Dec 14 17:16:07.040: INFO: Got endpoints: latency-svc-7cgb2 [744.16682ms] -Dec 14 17:16:07.056: INFO: Created: latency-svc-8lt4q -Dec 14 17:16:07.092: INFO: Got endpoints: latency-svc-t5plk [750.067526ms] -Dec 14 17:16:07.110: INFO: Created: latency-svc-q4gbc -Dec 14 17:16:07.146: INFO: Got endpoints: latency-svc-wc9wl [744.588083ms] -Dec 14 17:16:07.160: INFO: Created: latency-svc-vlrnh -Dec 14 17:16:07.196: INFO: Got endpoints: latency-svc-88vpx [756.5254ms] -Dec 14 17:16:07.216: INFO: Created: latency-svc-8dpss -Dec 14 17:16:07.239: INFO: Got endpoints: latency-svc-s82xq [743.022723ms] -Dec 14 17:16:07.256: INFO: Created: latency-svc-pxqfv -Dec 14 17:16:07.290: INFO: Got endpoints: latency-svc-9j7g9 [750.692442ms] -Dec 14 17:16:07.315: INFO: Created: latency-svc-n9qnp -Dec 14 17:16:07.340: INFO: Got endpoints: latency-svc-5jjmr [749.899751ms] -Dec 14 17:16:07.359: INFO: Created: latency-svc-68mm6 -Dec 14 17:16:07.393: INFO: Got endpoints: latency-svc-8zmzg [752.146478ms] -Dec 14 17:16:07.412: INFO: Created: latency-svc-98q9n -Dec 14 17:16:07.440: INFO: Got endpoints: latency-svc-jr26z [749.506755ms] -Dec 14 17:16:07.458: INFO: Created: latency-svc-g4p92 -Dec 14 17:16:07.491: INFO: Got endpoints: latency-svc-z6kb2 [750.740785ms] -Dec 14 17:16:07.508: INFO: Created: latency-svc-2ktd2 -Dec 14 17:16:07.541: INFO: Got endpoints: latency-svc-cmnsh [753.13947ms] -Dec 14 17:16:07.556: INFO: Created: latency-svc-d7wtw -Dec 14 17:16:07.590: INFO: Got endpoints: latency-svc-56mq6 [750.121135ms] -Dec 14 17:16:07.606: INFO: Created: latency-svc-tf9nq -Dec 14 17:16:07.646: INFO: Got endpoints: latency-svc-2xvvr [748.11265ms] -Dec 14 17:16:07.685: INFO: Created: latency-svc-4ccp4 -Dec 14 17:16:07.689: INFO: Got endpoints: latency-svc-pks7q [748.744933ms] -Dec 14 17:16:07.741: INFO: Got endpoints: latency-svc-xfj2x [750.461963ms] -Dec 14 17:16:07.762: INFO: Created: latency-svc-cs7b8 -Dec 14 17:16:07.768: INFO: Created: latency-svc-db5ws -Dec 14 17:16:07.812: INFO: Got endpoints: latency-svc-8lt4q [771.7579ms] -Dec 14 17:16:07.857: INFO: Got endpoints: latency-svc-q4gbc [765.06147ms] -Dec 14 17:16:07.895: INFO: Created: latency-svc-6qscx -Dec 14 17:16:07.911: INFO: Created: latency-svc-54s9m -Dec 14 17:16:07.932: INFO: Got endpoints: latency-svc-vlrnh [786.227668ms] -Dec 14 17:16:07.945: INFO: Created: latency-svc-889cc -Dec 14 17:16:07.976: INFO: Got endpoints: latency-svc-8dpss [779.433421ms] -Dec 14 17:16:08.013: INFO: Got endpoints: latency-svc-pxqfv [773.339119ms] -Dec 14 17:16:08.035: INFO: Created: latency-svc-b2vgf -Dec 14 17:16:08.057: INFO: Created: latency-svc-c6zlk -Dec 14 17:16:08.104: INFO: Got endpoints: latency-svc-68mm6 [764.25947ms] -Dec 14 17:16:08.121: INFO: Created: latency-svc-pqbjk -Dec 14 17:16:08.180: INFO: Got endpoints: latency-svc-98q9n [786.326179ms] -Dec 14 17:16:08.222: INFO: Got endpoints: latency-svc-n9qnp [931.482307ms] -Dec 14 17:16:08.222: INFO: Got endpoints: latency-svc-g4p92 [782.124819ms] -Dec 14 17:16:08.249: INFO: Got endpoints: latency-svc-2ktd2 [757.148494ms] -Dec 14 17:16:08.252: INFO: Created: latency-svc-xhvt5 -Dec 14 17:16:08.260: INFO: Created: latency-svc-hcr5q -Dec 14 17:16:08.269: INFO: Created: latency-svc-5m442 -Dec 14 17:16:08.306: INFO: Got endpoints: latency-svc-d7wtw [765.192545ms] -Dec 14 17:16:08.313: INFO: Created: latency-svc-pcn6z -Dec 14 17:16:08.326: INFO: Created: latency-svc-xvjgf -Dec 14 17:16:08.347: INFO: Got endpoints: latency-svc-tf9nq [756.855239ms] -Dec 14 17:16:08.404: INFO: Got endpoints: latency-svc-4ccp4 [757.992513ms] -Dec 14 17:16:08.406: INFO: Created: latency-svc-hsc22 -Dec 14 17:16:08.438: INFO: Created: latency-svc-5krnr -Dec 14 17:16:08.441: INFO: Got endpoints: latency-svc-cs7b8 [751.566716ms] -Dec 14 17:16:08.458: INFO: Created: latency-svc-chsb9 -Dec 14 17:16:08.489: INFO: Got endpoints: latency-svc-db5ws [748.401585ms] -Dec 14 17:16:08.516: INFO: Created: latency-svc-5t67z -Dec 14 17:16:08.539: INFO: Got endpoints: latency-svc-6qscx [726.832839ms] -Dec 14 17:16:08.555: INFO: Created: latency-svc-gl4jl -Dec 14 17:16:08.591: INFO: Got endpoints: latency-svc-54s9m [734.575075ms] -Dec 14 17:16:08.610: INFO: Created: latency-svc-swf6w -Dec 14 17:16:08.640: INFO: Got endpoints: latency-svc-889cc [707.882394ms] -Dec 14 17:16:08.662: INFO: Created: latency-svc-79hmd -Dec 14 17:16:08.689: INFO: Got endpoints: latency-svc-b2vgf [713.327205ms] -Dec 14 17:16:08.712: INFO: Created: latency-svc-7g9dj -Dec 14 17:16:08.738: INFO: Got endpoints: latency-svc-c6zlk [724.981018ms] -Dec 14 17:16:08.753: INFO: Created: latency-svc-l9lvw -Dec 14 17:16:08.790: INFO: Got endpoints: latency-svc-pqbjk [685.342031ms] -Dec 14 17:16:08.802: INFO: Created: latency-svc-pzfcw -Dec 14 17:16:08.841: INFO: Got endpoints: latency-svc-xhvt5 [661.04851ms] -Dec 14 17:16:08.862: INFO: Created: latency-svc-mwskq -Dec 14 17:16:08.890: INFO: Got endpoints: latency-svc-hcr5q [668.502784ms] -Dec 14 17:16:08.917: INFO: Created: latency-svc-87v8c -Dec 14 17:16:08.943: INFO: Got endpoints: latency-svc-5m442 [720.144517ms] -Dec 14 17:16:08.974: INFO: Created: latency-svc-7l6hw -Dec 14 17:16:08.995: INFO: Got endpoints: latency-svc-pcn6z [743.86906ms] -Dec 14 17:16:09.013: INFO: Created: latency-svc-wm8tz -Dec 14 17:16:09.043: INFO: Got endpoints: latency-svc-xvjgf [733.226417ms] -Dec 14 17:16:09.064: INFO: Created: latency-svc-7426c -Dec 14 17:16:09.096: INFO: Got endpoints: latency-svc-hsc22 [723.16495ms] -Dec 14 17:16:09.118: INFO: Created: latency-svc-q58nh -Dec 14 17:16:09.145: INFO: Got endpoints: latency-svc-5krnr [739.691885ms] -Dec 14 17:16:09.178: INFO: Created: latency-svc-rbrvp -Dec 14 17:16:09.234: INFO: Got endpoints: latency-svc-chsb9 [791.337642ms] -Dec 14 17:16:09.248: INFO: Got endpoints: latency-svc-5t67z [745.69114ms] -Dec 14 17:16:09.288: INFO: Created: latency-svc-m99rw -Dec 14 17:16:09.311: INFO: Got endpoints: latency-svc-gl4jl [771.193668ms] -Dec 14 17:16:09.318: INFO: Created: latency-svc-r2fjf -Dec 14 17:16:09.345: INFO: Created: latency-svc-t2mbj -Dec 14 17:16:09.350: INFO: Got endpoints: latency-svc-swf6w [758.875675ms] -Dec 14 17:16:09.366: INFO: Created: latency-svc-w4x67 -Dec 14 17:16:09.390: INFO: Got endpoints: latency-svc-79hmd [749.373759ms] -Dec 14 17:16:09.439: INFO: Created: latency-svc-bdbzn -Dec 14 17:16:09.445: INFO: Got endpoints: latency-svc-7g9dj [756.067844ms] -Dec 14 17:16:09.462: INFO: Created: latency-svc-bsblm -Dec 14 17:16:09.500: INFO: Got endpoints: latency-svc-l9lvw [762.16168ms] -Dec 14 17:16:09.526: INFO: Created: latency-svc-jwx52 -Dec 14 17:16:09.550: INFO: Got endpoints: latency-svc-pzfcw [759.627201ms] -Dec 14 17:16:09.570: INFO: Created: latency-svc-2gxks -Dec 14 17:16:09.593: INFO: Got endpoints: latency-svc-mwskq [752.1058ms] -Dec 14 17:16:09.606: INFO: Created: latency-svc-vwr9g -Dec 14 17:16:09.642: INFO: Got endpoints: latency-svc-87v8c [751.667485ms] -Dec 14 17:16:09.656: INFO: Created: latency-svc-m8brp -Dec 14 17:16:09.694: INFO: Got endpoints: latency-svc-7l6hw [751.210027ms] -Dec 14 17:16:09.711: INFO: Created: latency-svc-5snpr -Dec 14 17:16:09.740: INFO: Got endpoints: latency-svc-wm8tz [745.206213ms] -Dec 14 17:16:09.775: INFO: Created: latency-svc-jrjmt -Dec 14 17:16:09.802: INFO: Got endpoints: latency-svc-7426c [759.234702ms] -Dec 14 17:16:09.845: INFO: Got endpoints: latency-svc-q58nh [748.353189ms] -Dec 14 17:16:09.883: INFO: Created: latency-svc-w5l7k -Dec 14 17:16:09.895: INFO: Got endpoints: latency-svc-rbrvp [750.205498ms] -Dec 14 17:16:09.906: INFO: Created: latency-svc-jhhpb -Dec 14 17:16:09.923: INFO: Created: latency-svc-8pnkx -Dec 14 17:16:09.942: INFO: Got endpoints: latency-svc-m99rw [707.476595ms] -Dec 14 17:16:09.963: INFO: Created: latency-svc-hhvql -Dec 14 17:16:10.004: INFO: Got endpoints: latency-svc-r2fjf [755.622399ms] -Dec 14 17:16:10.034: INFO: Created: latency-svc-29jd4 -Dec 14 17:16:10.043: INFO: Got endpoints: latency-svc-t2mbj [732.385935ms] -Dec 14 17:16:10.067: INFO: Created: latency-svc-ntd4n -Dec 14 17:16:10.091: INFO: Got endpoints: latency-svc-w4x67 [740.24763ms] -Dec 14 17:16:10.106: INFO: Created: latency-svc-z8zx6 -Dec 14 17:16:10.145: INFO: Got endpoints: latency-svc-bdbzn [754.929511ms] -Dec 14 17:16:10.165: INFO: Created: latency-svc-vbb4n -Dec 14 17:16:10.200: INFO: Got endpoints: latency-svc-bsblm [754.267231ms] -Dec 14 17:16:10.218: INFO: Created: latency-svc-nhcqj -Dec 14 17:16:10.240: INFO: Got endpoints: latency-svc-jwx52 [739.198718ms] -Dec 14 17:16:10.253: INFO: Created: latency-svc-sl6j8 -Dec 14 17:16:10.291: INFO: Got endpoints: latency-svc-2gxks [740.911638ms] -Dec 14 17:16:10.303: INFO: Created: latency-svc-sn262 -Dec 14 17:16:10.341: INFO: Got endpoints: latency-svc-vwr9g [747.761301ms] -Dec 14 17:16:10.364: INFO: Created: latency-svc-gglvr -Dec 14 17:16:10.391: INFO: Got endpoints: latency-svc-m8brp [748.78858ms] -Dec 14 17:16:10.406: INFO: Created: latency-svc-pp26q -Dec 14 17:16:10.438: INFO: Got endpoints: latency-svc-5snpr [743.801082ms] -Dec 14 17:16:10.456: INFO: Created: latency-svc-2xzfw -Dec 14 17:16:10.489: INFO: Got endpoints: latency-svc-jrjmt [748.813722ms] -Dec 14 17:16:10.515: INFO: Created: latency-svc-fh6tz -Dec 14 17:16:10.546: INFO: Got endpoints: latency-svc-w5l7k [743.821643ms] -Dec 14 17:16:10.562: INFO: Created: latency-svc-q6w8r -Dec 14 17:16:10.590: INFO: Got endpoints: latency-svc-jhhpb [744.90311ms] -Dec 14 17:16:10.611: INFO: Created: latency-svc-wkh79 -Dec 14 17:16:10.640: INFO: Got endpoints: latency-svc-8pnkx [745.236967ms] -Dec 14 17:16:10.672: INFO: Created: latency-svc-kq8lz -Dec 14 17:16:10.690: INFO: Got endpoints: latency-svc-hhvql [747.853666ms] -Dec 14 17:16:10.702: INFO: Created: latency-svc-dz6jx -Dec 14 17:16:10.741: INFO: Got endpoints: latency-svc-29jd4 [736.818536ms] -Dec 14 17:16:10.756: INFO: Created: latency-svc-pqphm -Dec 14 17:16:10.797: INFO: Got endpoints: latency-svc-ntd4n [753.356684ms] -Dec 14 17:16:10.841: INFO: Got endpoints: latency-svc-z8zx6 [750.221257ms] -Dec 14 17:16:10.896: INFO: Got endpoints: latency-svc-vbb4n [751.164319ms] -Dec 14 17:16:10.941: INFO: Got endpoints: latency-svc-nhcqj [741.734122ms] -Dec 14 17:16:10.994: INFO: Got endpoints: latency-svc-sl6j8 [754.540575ms] -Dec 14 17:16:11.042: INFO: Got endpoints: latency-svc-sn262 [751.264467ms] -Dec 14 17:16:11.093: INFO: Got endpoints: latency-svc-gglvr [752.072589ms] -Dec 14 17:16:11.149: INFO: Got endpoints: latency-svc-pp26q [757.885809ms] -Dec 14 17:16:11.195: INFO: Got endpoints: latency-svc-2xzfw [757.382418ms] -Dec 14 17:16:11.239: INFO: Got endpoints: latency-svc-fh6tz [750.00636ms] -Dec 14 17:16:11.292: INFO: Got endpoints: latency-svc-q6w8r [745.125347ms] -Dec 14 17:16:11.340: INFO: Got endpoints: latency-svc-wkh79 [750.516841ms] -Dec 14 17:16:11.391: INFO: Got endpoints: latency-svc-kq8lz [750.885554ms] -Dec 14 17:16:11.441: INFO: Got endpoints: latency-svc-dz6jx [751.218228ms] -Dec 14 17:16:11.494: INFO: Got endpoints: latency-svc-pqphm [752.809407ms] -Dec 14 17:16:11.494: INFO: Latencies: [31.396245ms 141.208819ms 141.994896ms 163.230592ms 165.788645ms 182.137895ms 183.797748ms 196.846151ms 203.596618ms 221.701232ms 234.213126ms 240.271249ms 246.711469ms 253.527213ms 255.97453ms 257.992284ms 261.528179ms 262.079176ms 273.54668ms 281.270241ms 281.859048ms 283.741899ms 284.665018ms 287.891065ms 297.766123ms 299.208111ms 304.480559ms 309.272446ms 309.409449ms 310.229673ms 317.421903ms 318.867562ms 319.346279ms 329.47754ms 332.274215ms 336.006585ms 336.170935ms 336.421299ms 337.007557ms 341.811913ms 342.52653ms 345.118266ms 350.53158ms 353.648852ms 354.620515ms 355.575227ms 362.204865ms 365.509196ms 376.852681ms 382.685597ms 394.641869ms 417.611771ms 464.683146ms 480.530728ms 508.218847ms 525.76334ms 574.349617ms 580.957445ms 617.396396ms 653.692434ms 661.04851ms 668.502784ms 678.87874ms 681.836381ms 685.342031ms 700.445312ms 703.945132ms 706.774845ms 707.476595ms 707.882394ms 713.327205ms 715.473646ms 715.563427ms 717.573346ms 720.144517ms 723.16495ms 724.981018ms 725.704781ms 726.069733ms 726.832839ms 732.385935ms 733.226417ms 734.362717ms 734.575075ms 736.818536ms 739.198718ms 739.691885ms 740.24763ms 740.911638ms 741.734122ms 742.61698ms 743.022723ms 743.416321ms 743.801082ms 743.821643ms 743.86906ms 743.928118ms 743.985635ms 744.16682ms 744.226084ms 744.555911ms 744.588083ms 744.640515ms 744.678451ms 744.700531ms 744.90311ms 744.925734ms 745.125347ms 745.206213ms 745.236967ms 745.474333ms 745.64416ms 745.69114ms 746.199526ms 746.348404ms 747.051974ms 747.40927ms 747.761301ms 747.853666ms 748.11265ms 748.353189ms 748.388158ms 748.401585ms 748.536241ms 748.5896ms 748.744933ms 748.78858ms 748.813722ms 749.373759ms 749.506755ms 749.899751ms 750.00636ms 750.067526ms 750.121135ms 750.205498ms 750.221257ms 750.253503ms 750.461963ms 750.516841ms 750.692442ms 750.740785ms 750.885554ms 750.960083ms 751.164319ms 751.198596ms 751.210027ms 751.218228ms 751.264467ms 751.566716ms 751.667485ms 752.072589ms 752.1058ms 752.146478ms 752.809407ms 752.865812ms 753.13947ms 753.171167ms 753.313339ms 753.356684ms 753.495156ms 753.535874ms 753.905958ms 754.138026ms 754.267231ms 754.540575ms 754.66886ms 754.929511ms 755.05949ms 755.622399ms 756.067844ms 756.148074ms 756.5254ms 756.855239ms 757.148494ms 757.382418ms 757.885809ms 757.992513ms 758.251774ms 758.569214ms 758.875675ms 759.234702ms 759.466289ms 759.627201ms 762.16168ms 764.25947ms 765.06147ms 765.192545ms 768.221977ms 771.193668ms 771.7579ms 773.339119ms 779.433421ms 781.110117ms 782.124819ms 786.227668ms 786.326179ms 791.337642ms 793.324979ms 823.091477ms 931.482307ms] -Dec 14 17:16:11.494: INFO: 50 %ile: 744.555911ms -Dec 14 17:16:11.494: INFO: 90 %ile: 759.234702ms -Dec 14 17:16:11.494: INFO: 99 %ile: 823.091477ms -Dec 14 17:16:11.494: INFO: Total sample count: 200 -[AfterEach] [sig-network] Service endpoints latency +STEP: Creating a kubernetes client 02/25/23 11:43:30.014 +Feb 25 11:43:30.014: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename namespaces 02/25/23 11:43:30.019 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:43:30.073 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:43:30.08 +[It] should ensure that all services are removed when a namespace is deleted [Conformance] + test/e2e/apimachinery/namespace.go:250 +STEP: Creating a test namespace 02/25/23 11:43:30.084 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:43:30.19 +STEP: Creating a service in the namespace 02/25/23 11:43:30.201 +STEP: Deleting the namespace 02/25/23 11:43:30.225 +STEP: Waiting for the namespace to be removed. 02/25/23 11:43:30.241 +STEP: Recreating the namespace 02/25/23 11:43:36.25 +STEP: Verifying there is no service in the namespace 02/25/23 11:43:36.282 +[AfterEach] [sig-api-machinery] Namespaces [Serial] test/e2e/framework/framework.go:187 -Dec 14 17:16:11.494: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "svc-latency-4007" for this suite. 12/14/22 17:16:11.507 -{"msg":"PASSED [sig-network] Service endpoints latency should not be very high [Conformance]","completed":360,"skipped":6649,"failed":0} +Feb 25 11:43:36.287: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "namespaces-1616" for this suite. 02/25/23 11:43:36.304 +STEP: Destroying namespace "nsdeletetest-3035" for this suite. 02/25/23 11:43:36.316 +Feb 25 11:43:36.322: INFO: Namespace nsdeletetest-3035 was already deleted +STEP: Destroying namespace "nsdeletetest-9683" for this suite. 02/25/23 11:43:36.322 +{"msg":"PASSED [sig-api-machinery] Namespaces [Serial] should ensure that all services are removed when a namespace is deleted [Conformance]","completed":360,"skipped":6677,"failed":0} ------------------------------ -• [SLOW TEST] [10.935 seconds] -[sig-network] Service endpoints latency -test/e2e/network/common/framework.go:23 - should not be very high [Conformance] - test/e2e/network/service_latency.go:59 +• [SLOW TEST] [6.319 seconds] +[sig-api-machinery] Namespaces [Serial] +test/e2e/apimachinery/framework.go:23 + should ensure that all services are removed when a namespace is deleted [Conformance] + test/e2e/apimachinery/namespace.go:250 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] Service endpoints latency + [BeforeEach] [sig-api-machinery] Namespaces [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:16:00.581 - Dec 14 17:16:00.582: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename svc-latency 12/14/22 17:16:00.588 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:16:00.65 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:16:00.654 - [It] should not be very high [Conformance] - test/e2e/network/service_latency.go:59 - Dec 14 17:16:00.657: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: creating replication controller svc-latency-rc in namespace svc-latency-4007 12/14/22 17:16:00.66 - I1214 17:16:00.670883 14 runners.go:193] Created replication controller with name: svc-latency-rc, namespace: svc-latency-4007, replica count: 1 - I1214 17:16:01.722077 14 runners.go:193] svc-latency-rc Pods: 1 out of 1 created, 0 running, 1 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - I1214 17:16:02.722385 14 runners.go:193] svc-latency-rc Pods: 1 out of 1 created, 1 running, 0 pending, 0 waiting, 0 inactive, 0 terminating, 0 unknown, 0 runningButNotReady - Dec 14 17:16:02.837: INFO: Created: latency-svc-qwk5l - Dec 14 17:16:02.874: INFO: Got endpoints: latency-svc-qwk5l [51.950228ms] - Dec 14 17:16:02.895: INFO: Created: latency-svc-xns98 - Dec 14 17:16:02.906: INFO: Got endpoints: latency-svc-xns98 [31.396245ms] - Dec 14 17:16:03.027: INFO: Created: latency-svc-kk4xc - Dec 14 17:16:03.028: INFO: Created: latency-svc-vq4vn - Dec 14 17:16:03.028: INFO: Created: latency-svc-hqgx4 - Dec 14 17:16:03.028: INFO: Created: latency-svc-lp6rw - Dec 14 17:16:03.029: INFO: Created: latency-svc-8lzd5 - Dec 14 17:16:03.030: INFO: Created: latency-svc-7dc67 - Dec 14 17:16:03.035: INFO: Created: latency-svc-8t5c9 - Dec 14 17:16:03.036: INFO: Created: latency-svc-nzcxn - Dec 14 17:16:03.036: INFO: Created: latency-svc-vrhnk - Dec 14 17:16:03.036: INFO: Created: latency-svc-92vlc - Dec 14 17:16:03.041: INFO: Got endpoints: latency-svc-7dc67 [165.788645ms] - Dec 14 17:16:03.055: INFO: Created: latency-svc-thjr2 - Dec 14 17:16:03.056: INFO: Created: latency-svc-wrmr9 - Dec 14 17:16:03.063: INFO: Created: latency-svc-kjkrj - Dec 14 17:16:03.133: INFO: Got endpoints: latency-svc-thjr2 [257.992284ms] - Dec 14 17:16:03.137: INFO: Got endpoints: latency-svc-wrmr9 [262.079176ms] - Dec 14 17:16:03.142: INFO: Created: latency-svc-7b7lt - Dec 14 17:16:03.143: INFO: Created: latency-svc-4pntf - Dec 14 17:16:03.146: INFO: Got endpoints: latency-svc-8lzd5 [240.271249ms] - Dec 14 17:16:03.157: INFO: Got endpoints: latency-svc-kk4xc [281.270241ms] - Dec 14 17:16:03.163: INFO: Got endpoints: latency-svc-lp6rw [287.891065ms] - Dec 14 17:16:03.167: INFO: Created: latency-svc-l8d8b - Dec 14 17:16:03.186: INFO: Got endpoints: latency-svc-hqgx4 [309.272446ms] - Dec 14 17:16:03.193: INFO: Got endpoints: latency-svc-vq4vn [317.421903ms] - Dec 14 17:16:03.194: INFO: Got endpoints: latency-svc-92vlc [318.867562ms] - Dec 14 17:16:03.195: INFO: Got endpoints: latency-svc-8t5c9 [319.346279ms] - Dec 14 17:16:03.202: INFO: Created: latency-svc-5ljws - Dec 14 17:16:03.213: INFO: Got endpoints: latency-svc-nzcxn [337.007557ms] - Dec 14 17:16:03.231: INFO: Got endpoints: latency-svc-vrhnk [354.620515ms] - Dec 14 17:16:03.231: INFO: Got endpoints: latency-svc-4pntf [355.575227ms] - Dec 14 17:16:03.238: INFO: Created: latency-svc-4qm9z - Dec 14 17:16:03.239: INFO: Created: latency-svc-dzpwx - Dec 14 17:16:03.259: INFO: Got endpoints: latency-svc-7b7lt [382.685597ms] - Dec 14 17:16:03.270: INFO: Got endpoints: latency-svc-l8d8b [394.641869ms] - Dec 14 17:16:03.274: INFO: Got endpoints: latency-svc-5ljws [141.208819ms] - Dec 14 17:16:03.275: INFO: Got endpoints: latency-svc-kjkrj [234.213126ms] - Dec 14 17:16:03.279: INFO: Got endpoints: latency-svc-4qm9z [141.994896ms] - Dec 14 17:16:03.288: INFO: Created: latency-svc-p58cv - Dec 14 17:16:03.309: INFO: Got endpoints: latency-svc-dzpwx [163.230592ms] - Dec 14 17:16:03.341: INFO: Got endpoints: latency-svc-p58cv [183.797748ms] - Dec 14 17:16:03.354: INFO: Created: latency-svc-4hg2p - Dec 14 17:16:03.355: INFO: Created: latency-svc-2rrjd - Dec 14 17:16:03.360: INFO: Got endpoints: latency-svc-2rrjd [196.846151ms] - Dec 14 17:16:03.368: INFO: Got endpoints: latency-svc-4hg2p [182.137895ms] - Dec 14 17:16:03.385: INFO: Created: latency-svc-8rwz9 - Dec 14 17:16:03.398: INFO: Got endpoints: latency-svc-8rwz9 [203.596618ms] - Dec 14 17:16:03.408: INFO: Created: latency-svc-jrjlh - Dec 14 17:16:03.415: INFO: Got endpoints: latency-svc-jrjlh [221.701232ms] - Dec 14 17:16:03.418: INFO: Created: latency-svc-tkscz - Dec 14 17:16:03.434: INFO: Created: latency-svc-mws6h - Dec 14 17:16:03.442: INFO: Got endpoints: latency-svc-tkscz [246.711469ms] - Dec 14 17:16:03.448: INFO: Created: latency-svc-c8hw7 - Dec 14 17:16:03.461: INFO: Created: latency-svc-2rkf8 - Dec 14 17:16:03.474: INFO: Got endpoints: latency-svc-mws6h [261.528179ms] - Dec 14 17:16:03.488: INFO: Got endpoints: latency-svc-c8hw7 [255.97453ms] - Dec 14 17:16:03.494: INFO: Created: latency-svc-txpz8 - Dec 14 17:16:03.515: INFO: Got endpoints: latency-svc-2rkf8 [283.741899ms] - Dec 14 17:16:03.524: INFO: Got endpoints: latency-svc-txpz8 [253.527213ms] - Dec 14 17:16:03.557: INFO: Created: latency-svc-c98qw - Dec 14 17:16:03.574: INFO: Got endpoints: latency-svc-c98qw [299.208111ms] - Dec 14 17:16:03.577: INFO: Created: latency-svc-bvbqk - Dec 14 17:16:03.604: INFO: Got endpoints: latency-svc-bvbqk [329.47754ms] - Dec 14 17:16:03.620: INFO: Created: latency-svc-wnvrr - Dec 14 17:16:03.645: INFO: Got endpoints: latency-svc-wnvrr [365.509196ms] - Dec 14 17:16:03.652: INFO: Created: latency-svc-5n8rf - Dec 14 17:16:03.665: INFO: Created: latency-svc-nxbr6 - Dec 14 17:16:03.672: INFO: Got endpoints: latency-svc-5n8rf [362.204865ms] - Dec 14 17:16:03.692: INFO: Got endpoints: latency-svc-nxbr6 [350.53158ms] - Dec 14 17:16:03.699: INFO: Created: latency-svc-ptfwd - Dec 14 17:16:03.705: INFO: Got endpoints: latency-svc-ptfwd [345.118266ms] - Dec 14 17:16:03.709: INFO: Created: latency-svc-78ksf - Dec 14 17:16:03.722: INFO: Created: latency-svc-7k7ds - Dec 14 17:16:03.722: INFO: Got endpoints: latency-svc-78ksf [353.648852ms] - Dec 14 17:16:03.730: INFO: Created: latency-svc-mczqt - Dec 14 17:16:03.730: INFO: Got endpoints: latency-svc-7k7ds [332.274215ms] - Dec 14 17:16:03.742: INFO: Created: latency-svc-bkzp6 - Dec 14 17:16:03.748: INFO: Created: latency-svc-hdcr6 - Dec 14 17:16:03.751: INFO: Created: latency-svc-76vtw - Dec 14 17:16:03.757: INFO: Got endpoints: latency-svc-mczqt [341.811913ms] - Dec 14 17:16:03.783: INFO: Created: latency-svc-zmwmw - Dec 14 17:16:03.784: INFO: Got endpoints: latency-svc-hdcr6 [342.52653ms] - Dec 14 17:16:03.785: INFO: Got endpoints: latency-svc-76vtw [310.229673ms] - Dec 14 17:16:03.785: INFO: Got endpoints: latency-svc-bkzp6 [525.76334ms] - Dec 14 17:16:03.794: INFO: Created: latency-svc-4sjvg - Dec 14 17:16:03.823: INFO: Created: latency-svc-vcnph - Dec 14 17:16:03.824: INFO: Got endpoints: latency-svc-zmwmw [336.006585ms] - Dec 14 17:16:03.824: INFO: Got endpoints: latency-svc-4sjvg [309.409449ms] - Dec 14 17:16:03.860: INFO: Got endpoints: latency-svc-vcnph [336.170935ms] - Dec 14 17:16:03.868: INFO: Created: latency-svc-8ztnf - Dec 14 17:16:03.877: INFO: Created: latency-svc-nllnl - Dec 14 17:16:03.878: INFO: Got endpoints: latency-svc-8ztnf [304.480559ms] - Dec 14 17:16:03.889: INFO: Got endpoints: latency-svc-nllnl [284.665018ms] - Dec 14 17:16:03.900: INFO: Created: latency-svc-7m8gx - Dec 14 17:16:03.927: INFO: Got endpoints: latency-svc-7m8gx [281.859048ms] - Dec 14 17:16:03.940: INFO: Created: latency-svc-kx4x5 - Dec 14 17:16:03.940: INFO: Created: latency-svc-5trjp - Dec 14 17:16:03.946: INFO: Got endpoints: latency-svc-kx4x5 [273.54668ms] - Dec 14 17:16:03.969: INFO: Created: latency-svc-9ms4v - Dec 14 17:16:03.990: INFO: Got endpoints: latency-svc-5trjp [297.766123ms] - Dec 14 17:16:03.996: INFO: Created: latency-svc-c2mxz - Dec 14 17:16:04.004: INFO: Created: latency-svc-qxbsp - Dec 14 17:16:04.030: INFO: Created: latency-svc-cblft - Dec 14 17:16:04.042: INFO: Got endpoints: latency-svc-9ms4v [336.421299ms] - Dec 14 17:16:04.055: INFO: Created: latency-svc-r24vc - Dec 14 17:16:04.065: INFO: Created: latency-svc-2kdkx - Dec 14 17:16:04.076: INFO: Created: latency-svc-p7hwp - Dec 14 17:16:04.099: INFO: Got endpoints: latency-svc-qxbsp [376.852681ms] - Dec 14 17:16:04.112: INFO: Created: latency-svc-rf99f - Dec 14 17:16:04.130: INFO: Created: latency-svc-dmt2n - Dec 14 17:16:04.148: INFO: Got endpoints: latency-svc-c2mxz [417.611771ms] - Dec 14 17:16:04.149: INFO: Created: latency-svc-6qpnz - Dec 14 17:16:04.164: INFO: Created: latency-svc-z8vn5 - Dec 14 17:16:04.169: INFO: Created: latency-svc-5pmw7 - Dec 14 17:16:04.212: INFO: Created: latency-svc-k7cng - Dec 14 17:16:04.238: INFO: Got endpoints: latency-svc-cblft [480.530728ms] - Dec 14 17:16:04.249: INFO: Got endpoints: latency-svc-2kdkx [464.683146ms] - Dec 14 17:16:04.272: INFO: Created: latency-svc-xc5sn - Dec 14 17:16:04.276: INFO: Created: latency-svc-ls99t - Dec 14 17:16:04.293: INFO: Got endpoints: latency-svc-r24vc [508.218847ms] - Dec 14 17:16:04.365: INFO: Got endpoints: latency-svc-p7hwp [580.957445ms] - Dec 14 17:16:04.396: INFO: Created: latency-svc-r5frc - Dec 14 17:16:04.398: INFO: Got endpoints: latency-svc-rf99f [574.349617ms] - Dec 14 17:16:04.403: INFO: Created: latency-svc-w8jk6 - Dec 14 17:16:04.412: INFO: Created: latency-svc-xhd8p - Dec 14 17:16:04.442: INFO: Got endpoints: latency-svc-dmt2n [617.396396ms] - Dec 14 17:16:04.459: INFO: Created: latency-svc-c4tbm - Dec 14 17:16:04.474: INFO: Created: latency-svc-mvf9p - Dec 14 17:16:04.480: INFO: Created: latency-svc-mjl22 - Dec 14 17:16:04.504: INFO: Created: latency-svc-nzcsn - Dec 14 17:16:04.514: INFO: Got endpoints: latency-svc-6qpnz [653.692434ms] - Dec 14 17:16:04.525: INFO: Created: latency-svc-w4dkn - Dec 14 17:16:04.533: INFO: Created: latency-svc-kndgt - Dec 14 17:16:04.557: INFO: Got endpoints: latency-svc-z8vn5 [678.87874ms] - Dec 14 17:16:04.566: INFO: Created: latency-svc-5q9qg - Dec 14 17:16:04.580: INFO: Created: latency-svc-vbcnl - Dec 14 17:16:04.596: INFO: Got endpoints: latency-svc-5pmw7 [706.774845ms] - Dec 14 17:16:04.618: INFO: Created: latency-svc-lr569 - Dec 14 17:16:04.643: INFO: Got endpoints: latency-svc-k7cng [715.473646ms] - Dec 14 17:16:04.660: INFO: Created: latency-svc-m2fd9 - Dec 14 17:16:04.695: INFO: Got endpoints: latency-svc-xc5sn [748.536241ms] - Dec 14 17:16:04.723: INFO: Created: latency-svc-gd89s - Dec 14 17:16:04.745: INFO: Got endpoints: latency-svc-ls99t [755.05949ms] - Dec 14 17:16:04.773: INFO: Created: latency-svc-bs2kc - Dec 14 17:16:04.793: INFO: Got endpoints: latency-svc-r5frc [750.960083ms] - Dec 14 17:16:04.812: INFO: Created: latency-svc-n6ml6 - Dec 14 17:16:04.843: INFO: Got endpoints: latency-svc-w8jk6 [744.640515ms] - Dec 14 17:16:04.867: INFO: Created: latency-svc-6bvgc - Dec 14 17:16:04.895: INFO: Got endpoints: latency-svc-xhd8p [747.051974ms] - Dec 14 17:16:04.917: INFO: Created: latency-svc-ndhn7 - Dec 14 17:16:04.942: INFO: Got endpoints: latency-svc-c4tbm [703.945132ms] - Dec 14 17:16:04.959: INFO: Created: latency-svc-bnrr5 - Dec 14 17:16:04.994: INFO: Got endpoints: latency-svc-mvf9p [744.555911ms] - Dec 14 17:16:05.012: INFO: Created: latency-svc-dvh9h - Dec 14 17:16:05.041: INFO: Got endpoints: latency-svc-mjl22 [747.40927ms] - Dec 14 17:16:05.061: INFO: Created: latency-svc-kwcr4 - Dec 14 17:16:05.190: INFO: Got endpoints: latency-svc-nzcsn [823.091477ms] - Dec 14 17:16:05.193: INFO: Got endpoints: latency-svc-w4dkn [793.324979ms] - Dec 14 17:16:05.196: INFO: Got endpoints: latency-svc-kndgt [754.138026ms] - Dec 14 17:16:05.215: INFO: Created: latency-svc-bxkhg - Dec 14 17:16:05.225: INFO: Created: latency-svc-8ldjj - Dec 14 17:16:05.231: INFO: Created: latency-svc-j6hbc - Dec 14 17:16:05.240: INFO: Got endpoints: latency-svc-5q9qg [725.704781ms] - Dec 14 17:16:05.257: INFO: Created: latency-svc-xj724 - Dec 14 17:16:05.292: INFO: Got endpoints: latency-svc-vbcnl [734.362717ms] - Dec 14 17:16:05.312: INFO: Created: latency-svc-fp5b7 - Dec 14 17:16:05.341: INFO: Got endpoints: latency-svc-lr569 [744.678451ms] - Dec 14 17:16:05.355: INFO: Created: latency-svc-r2cgx - Dec 14 17:16:05.394: INFO: Got endpoints: latency-svc-m2fd9 [751.198596ms] - Dec 14 17:16:05.407: INFO: Created: latency-svc-tnkxc - Dec 14 17:16:05.443: INFO: Got endpoints: latency-svc-gd89s [748.388158ms] - Dec 14 17:16:05.458: INFO: Created: latency-svc-9zqhg - Dec 14 17:16:05.491: INFO: Got endpoints: latency-svc-bs2kc [746.199526ms] - Dec 14 17:16:05.505: INFO: Created: latency-svc-nhjvs - Dec 14 17:16:05.548: INFO: Got endpoints: latency-svc-n6ml6 [754.66886ms] - Dec 14 17:16:05.561: INFO: Created: latency-svc-mz2wn - Dec 14 17:16:05.589: INFO: Got endpoints: latency-svc-6bvgc [744.925734ms] - Dec 14 17:16:05.616: INFO: Created: latency-svc-7p5f4 - Dec 14 17:16:05.655: INFO: Got endpoints: latency-svc-ndhn7 [759.466289ms] - Dec 14 17:16:05.675: INFO: Created: latency-svc-tvdm6 - Dec 14 17:16:05.695: INFO: Got endpoints: latency-svc-bnrr5 [753.495156ms] - Dec 14 17:16:05.714: INFO: Created: latency-svc-zhkkb - Dec 14 17:16:05.738: INFO: Got endpoints: latency-svc-dvh9h [743.928118ms] - Dec 14 17:16:05.785: INFO: Created: latency-svc-k8hwb - Dec 14 17:16:05.795: INFO: Got endpoints: latency-svc-kwcr4 [753.535874ms] - Dec 14 17:16:05.817: INFO: Created: latency-svc-6kp27 - Dec 14 17:16:05.874: INFO: Got endpoints: latency-svc-bxkhg [681.836381ms] - Dec 14 17:16:05.891: INFO: Got endpoints: latency-svc-8ldjj [700.445312ms] - Dec 14 17:16:05.962: INFO: Created: latency-svc-mcvgr - Dec 14 17:16:05.963: INFO: Created: latency-svc-vvsqf - Dec 14 17:16:05.964: INFO: Got endpoints: latency-svc-j6hbc [768.221977ms] - Dec 14 17:16:05.977: INFO: Created: latency-svc-kr2sk - Dec 14 17:16:05.996: INFO: Got endpoints: latency-svc-xj724 [756.148074ms] - Dec 14 17:16:06.034: INFO: Created: latency-svc-cpwzn - Dec 14 17:16:06.045: INFO: Got endpoints: latency-svc-fp5b7 [753.171167ms] - Dec 14 17:16:06.077: INFO: Created: latency-svc-9ljxp - Dec 14 17:16:06.122: INFO: Got endpoints: latency-svc-r2cgx [781.110117ms] - Dec 14 17:16:06.137: INFO: Created: latency-svc-ssjjf - Dec 14 17:16:06.153: INFO: Got endpoints: latency-svc-tnkxc [758.569214ms] - Dec 14 17:16:06.184: INFO: Created: latency-svc-xdsqz - Dec 14 17:16:06.197: INFO: Got endpoints: latency-svc-9zqhg [753.905958ms] - Dec 14 17:16:06.211: INFO: Created: latency-svc-fwkr7 - Dec 14 17:16:06.245: INFO: Got endpoints: latency-svc-nhjvs [753.313339ms] - Dec 14 17:16:06.259: INFO: Created: latency-svc-rtrc2 - Dec 14 17:16:06.296: INFO: Got endpoints: latency-svc-mz2wn [748.5896ms] - Dec 14 17:16:06.312: INFO: Created: latency-svc-7cgb2 - Dec 14 17:16:06.341: INFO: Got endpoints: latency-svc-7p5f4 [752.865812ms] - Dec 14 17:16:06.379: INFO: Created: latency-svc-t5plk - Dec 14 17:16:06.401: INFO: Got endpoints: latency-svc-tvdm6 [746.348404ms] - Dec 14 17:16:06.414: INFO: Created: latency-svc-wc9wl - Dec 14 17:16:06.439: INFO: Got endpoints: latency-svc-zhkkb [743.985635ms] - Dec 14 17:16:06.459: INFO: Created: latency-svc-88vpx - Dec 14 17:16:06.496: INFO: Got endpoints: latency-svc-k8hwb [758.251774ms] - Dec 14 17:16:06.511: INFO: Created: latency-svc-s82xq - Dec 14 17:16:06.539: INFO: Got endpoints: latency-svc-6kp27 [744.700531ms] - Dec 14 17:16:06.555: INFO: Created: latency-svc-9j7g9 - Dec 14 17:16:06.590: INFO: Got endpoints: latency-svc-vvsqf [715.563427ms] - Dec 14 17:16:06.607: INFO: Created: latency-svc-5jjmr - Dec 14 17:16:06.641: INFO: Got endpoints: latency-svc-mcvgr [750.253503ms] - Dec 14 17:16:06.660: INFO: Created: latency-svc-8zmzg - Dec 14 17:16:06.690: INFO: Got endpoints: latency-svc-kr2sk [726.069733ms] - Dec 14 17:16:06.703: INFO: Created: latency-svc-jr26z - Dec 14 17:16:06.741: INFO: Got endpoints: latency-svc-cpwzn [744.226084ms] - Dec 14 17:16:06.755: INFO: Created: latency-svc-z6kb2 - Dec 14 17:16:06.788: INFO: Got endpoints: latency-svc-9ljxp [742.61698ms] - Dec 14 17:16:06.801: INFO: Created: latency-svc-cmnsh - Dec 14 17:16:06.840: INFO: Got endpoints: latency-svc-ssjjf [717.573346ms] - Dec 14 17:16:06.853: INFO: Created: latency-svc-56mq6 - Dec 14 17:16:06.898: INFO: Got endpoints: latency-svc-xdsqz [745.64416ms] - Dec 14 17:16:06.915: INFO: Created: latency-svc-2xvvr - Dec 14 17:16:06.940: INFO: Got endpoints: latency-svc-fwkr7 [743.416321ms] - Dec 14 17:16:06.957: INFO: Created: latency-svc-pks7q - Dec 14 17:16:06.990: INFO: Got endpoints: latency-svc-rtrc2 [745.474333ms] - Dec 14 17:16:07.001: INFO: Created: latency-svc-xfj2x - Dec 14 17:16:07.040: INFO: Got endpoints: latency-svc-7cgb2 [744.16682ms] - Dec 14 17:16:07.056: INFO: Created: latency-svc-8lt4q - Dec 14 17:16:07.092: INFO: Got endpoints: latency-svc-t5plk [750.067526ms] - Dec 14 17:16:07.110: INFO: Created: latency-svc-q4gbc - Dec 14 17:16:07.146: INFO: Got endpoints: latency-svc-wc9wl [744.588083ms] - Dec 14 17:16:07.160: INFO: Created: latency-svc-vlrnh - Dec 14 17:16:07.196: INFO: Got endpoints: latency-svc-88vpx [756.5254ms] - Dec 14 17:16:07.216: INFO: Created: latency-svc-8dpss - Dec 14 17:16:07.239: INFO: Got endpoints: latency-svc-s82xq [743.022723ms] - Dec 14 17:16:07.256: INFO: Created: latency-svc-pxqfv - Dec 14 17:16:07.290: INFO: Got endpoints: latency-svc-9j7g9 [750.692442ms] - Dec 14 17:16:07.315: INFO: Created: latency-svc-n9qnp - Dec 14 17:16:07.340: INFO: Got endpoints: latency-svc-5jjmr [749.899751ms] - Dec 14 17:16:07.359: INFO: Created: latency-svc-68mm6 - Dec 14 17:16:07.393: INFO: Got endpoints: latency-svc-8zmzg [752.146478ms] - Dec 14 17:16:07.412: INFO: Created: latency-svc-98q9n - Dec 14 17:16:07.440: INFO: Got endpoints: latency-svc-jr26z [749.506755ms] - Dec 14 17:16:07.458: INFO: Created: latency-svc-g4p92 - Dec 14 17:16:07.491: INFO: Got endpoints: latency-svc-z6kb2 [750.740785ms] - Dec 14 17:16:07.508: INFO: Created: latency-svc-2ktd2 - Dec 14 17:16:07.541: INFO: Got endpoints: latency-svc-cmnsh [753.13947ms] - Dec 14 17:16:07.556: INFO: Created: latency-svc-d7wtw - Dec 14 17:16:07.590: INFO: Got endpoints: latency-svc-56mq6 [750.121135ms] - Dec 14 17:16:07.606: INFO: Created: latency-svc-tf9nq - Dec 14 17:16:07.646: INFO: Got endpoints: latency-svc-2xvvr [748.11265ms] - Dec 14 17:16:07.685: INFO: Created: latency-svc-4ccp4 - Dec 14 17:16:07.689: INFO: Got endpoints: latency-svc-pks7q [748.744933ms] - Dec 14 17:16:07.741: INFO: Got endpoints: latency-svc-xfj2x [750.461963ms] - Dec 14 17:16:07.762: INFO: Created: latency-svc-cs7b8 - Dec 14 17:16:07.768: INFO: Created: latency-svc-db5ws - Dec 14 17:16:07.812: INFO: Got endpoints: latency-svc-8lt4q [771.7579ms] - Dec 14 17:16:07.857: INFO: Got endpoints: latency-svc-q4gbc [765.06147ms] - Dec 14 17:16:07.895: INFO: Created: latency-svc-6qscx - Dec 14 17:16:07.911: INFO: Created: latency-svc-54s9m - Dec 14 17:16:07.932: INFO: Got endpoints: latency-svc-vlrnh [786.227668ms] - Dec 14 17:16:07.945: INFO: Created: latency-svc-889cc - Dec 14 17:16:07.976: INFO: Got endpoints: latency-svc-8dpss [779.433421ms] - Dec 14 17:16:08.013: INFO: Got endpoints: latency-svc-pxqfv [773.339119ms] - Dec 14 17:16:08.035: INFO: Created: latency-svc-b2vgf - Dec 14 17:16:08.057: INFO: Created: latency-svc-c6zlk - Dec 14 17:16:08.104: INFO: Got endpoints: latency-svc-68mm6 [764.25947ms] - Dec 14 17:16:08.121: INFO: Created: latency-svc-pqbjk - Dec 14 17:16:08.180: INFO: Got endpoints: latency-svc-98q9n [786.326179ms] - Dec 14 17:16:08.222: INFO: Got endpoints: latency-svc-n9qnp [931.482307ms] - Dec 14 17:16:08.222: INFO: Got endpoints: latency-svc-g4p92 [782.124819ms] - Dec 14 17:16:08.249: INFO: Got endpoints: latency-svc-2ktd2 [757.148494ms] - Dec 14 17:16:08.252: INFO: Created: latency-svc-xhvt5 - Dec 14 17:16:08.260: INFO: Created: latency-svc-hcr5q - Dec 14 17:16:08.269: INFO: Created: latency-svc-5m442 - Dec 14 17:16:08.306: INFO: Got endpoints: latency-svc-d7wtw [765.192545ms] - Dec 14 17:16:08.313: INFO: Created: latency-svc-pcn6z - Dec 14 17:16:08.326: INFO: Created: latency-svc-xvjgf - Dec 14 17:16:08.347: INFO: Got endpoints: latency-svc-tf9nq [756.855239ms] - Dec 14 17:16:08.404: INFO: Got endpoints: latency-svc-4ccp4 [757.992513ms] - Dec 14 17:16:08.406: INFO: Created: latency-svc-hsc22 - Dec 14 17:16:08.438: INFO: Created: latency-svc-5krnr - Dec 14 17:16:08.441: INFO: Got endpoints: latency-svc-cs7b8 [751.566716ms] - Dec 14 17:16:08.458: INFO: Created: latency-svc-chsb9 - Dec 14 17:16:08.489: INFO: Got endpoints: latency-svc-db5ws [748.401585ms] - Dec 14 17:16:08.516: INFO: Created: latency-svc-5t67z - Dec 14 17:16:08.539: INFO: Got endpoints: latency-svc-6qscx [726.832839ms] - Dec 14 17:16:08.555: INFO: Created: latency-svc-gl4jl - Dec 14 17:16:08.591: INFO: Got endpoints: latency-svc-54s9m [734.575075ms] - Dec 14 17:16:08.610: INFO: Created: latency-svc-swf6w - Dec 14 17:16:08.640: INFO: Got endpoints: latency-svc-889cc [707.882394ms] - Dec 14 17:16:08.662: INFO: Created: latency-svc-79hmd - Dec 14 17:16:08.689: INFO: Got endpoints: latency-svc-b2vgf [713.327205ms] - Dec 14 17:16:08.712: INFO: Created: latency-svc-7g9dj - Dec 14 17:16:08.738: INFO: Got endpoints: latency-svc-c6zlk [724.981018ms] - Dec 14 17:16:08.753: INFO: Created: latency-svc-l9lvw - Dec 14 17:16:08.790: INFO: Got endpoints: latency-svc-pqbjk [685.342031ms] - Dec 14 17:16:08.802: INFO: Created: latency-svc-pzfcw - Dec 14 17:16:08.841: INFO: Got endpoints: latency-svc-xhvt5 [661.04851ms] - Dec 14 17:16:08.862: INFO: Created: latency-svc-mwskq - Dec 14 17:16:08.890: INFO: Got endpoints: latency-svc-hcr5q [668.502784ms] - Dec 14 17:16:08.917: INFO: Created: latency-svc-87v8c - Dec 14 17:16:08.943: INFO: Got endpoints: latency-svc-5m442 [720.144517ms] - Dec 14 17:16:08.974: INFO: Created: latency-svc-7l6hw - Dec 14 17:16:08.995: INFO: Got endpoints: latency-svc-pcn6z [743.86906ms] - Dec 14 17:16:09.013: INFO: Created: latency-svc-wm8tz - Dec 14 17:16:09.043: INFO: Got endpoints: latency-svc-xvjgf [733.226417ms] - Dec 14 17:16:09.064: INFO: Created: latency-svc-7426c - Dec 14 17:16:09.096: INFO: Got endpoints: latency-svc-hsc22 [723.16495ms] - Dec 14 17:16:09.118: INFO: Created: latency-svc-q58nh - Dec 14 17:16:09.145: INFO: Got endpoints: latency-svc-5krnr [739.691885ms] - Dec 14 17:16:09.178: INFO: Created: latency-svc-rbrvp - Dec 14 17:16:09.234: INFO: Got endpoints: latency-svc-chsb9 [791.337642ms] - Dec 14 17:16:09.248: INFO: Got endpoints: latency-svc-5t67z [745.69114ms] - Dec 14 17:16:09.288: INFO: Created: latency-svc-m99rw - Dec 14 17:16:09.311: INFO: Got endpoints: latency-svc-gl4jl [771.193668ms] - Dec 14 17:16:09.318: INFO: Created: latency-svc-r2fjf - Dec 14 17:16:09.345: INFO: Created: latency-svc-t2mbj - Dec 14 17:16:09.350: INFO: Got endpoints: latency-svc-swf6w [758.875675ms] - Dec 14 17:16:09.366: INFO: Created: latency-svc-w4x67 - Dec 14 17:16:09.390: INFO: Got endpoints: latency-svc-79hmd [749.373759ms] - Dec 14 17:16:09.439: INFO: Created: latency-svc-bdbzn - Dec 14 17:16:09.445: INFO: Got endpoints: latency-svc-7g9dj [756.067844ms] - Dec 14 17:16:09.462: INFO: Created: latency-svc-bsblm - Dec 14 17:16:09.500: INFO: Got endpoints: latency-svc-l9lvw [762.16168ms] - Dec 14 17:16:09.526: INFO: Created: latency-svc-jwx52 - Dec 14 17:16:09.550: INFO: Got endpoints: latency-svc-pzfcw [759.627201ms] - Dec 14 17:16:09.570: INFO: Created: latency-svc-2gxks - Dec 14 17:16:09.593: INFO: Got endpoints: latency-svc-mwskq [752.1058ms] - Dec 14 17:16:09.606: INFO: Created: latency-svc-vwr9g - Dec 14 17:16:09.642: INFO: Got endpoints: latency-svc-87v8c [751.667485ms] - Dec 14 17:16:09.656: INFO: Created: latency-svc-m8brp - Dec 14 17:16:09.694: INFO: Got endpoints: latency-svc-7l6hw [751.210027ms] - Dec 14 17:16:09.711: INFO: Created: latency-svc-5snpr - Dec 14 17:16:09.740: INFO: Got endpoints: latency-svc-wm8tz [745.206213ms] - Dec 14 17:16:09.775: INFO: Created: latency-svc-jrjmt - Dec 14 17:16:09.802: INFO: Got endpoints: latency-svc-7426c [759.234702ms] - Dec 14 17:16:09.845: INFO: Got endpoints: latency-svc-q58nh [748.353189ms] - Dec 14 17:16:09.883: INFO: Created: latency-svc-w5l7k - Dec 14 17:16:09.895: INFO: Got endpoints: latency-svc-rbrvp [750.205498ms] - Dec 14 17:16:09.906: INFO: Created: latency-svc-jhhpb - Dec 14 17:16:09.923: INFO: Created: latency-svc-8pnkx - Dec 14 17:16:09.942: INFO: Got endpoints: latency-svc-m99rw [707.476595ms] - Dec 14 17:16:09.963: INFO: Created: latency-svc-hhvql - Dec 14 17:16:10.004: INFO: Got endpoints: latency-svc-r2fjf [755.622399ms] - Dec 14 17:16:10.034: INFO: Created: latency-svc-29jd4 - Dec 14 17:16:10.043: INFO: Got endpoints: latency-svc-t2mbj [732.385935ms] - Dec 14 17:16:10.067: INFO: Created: latency-svc-ntd4n - Dec 14 17:16:10.091: INFO: Got endpoints: latency-svc-w4x67 [740.24763ms] - Dec 14 17:16:10.106: INFO: Created: latency-svc-z8zx6 - Dec 14 17:16:10.145: INFO: Got endpoints: latency-svc-bdbzn [754.929511ms] - Dec 14 17:16:10.165: INFO: Created: latency-svc-vbb4n - Dec 14 17:16:10.200: INFO: Got endpoints: latency-svc-bsblm [754.267231ms] - Dec 14 17:16:10.218: INFO: Created: latency-svc-nhcqj - Dec 14 17:16:10.240: INFO: Got endpoints: latency-svc-jwx52 [739.198718ms] - Dec 14 17:16:10.253: INFO: Created: latency-svc-sl6j8 - Dec 14 17:16:10.291: INFO: Got endpoints: latency-svc-2gxks [740.911638ms] - Dec 14 17:16:10.303: INFO: Created: latency-svc-sn262 - Dec 14 17:16:10.341: INFO: Got endpoints: latency-svc-vwr9g [747.761301ms] - Dec 14 17:16:10.364: INFO: Created: latency-svc-gglvr - Dec 14 17:16:10.391: INFO: Got endpoints: latency-svc-m8brp [748.78858ms] - Dec 14 17:16:10.406: INFO: Created: latency-svc-pp26q - Dec 14 17:16:10.438: INFO: Got endpoints: latency-svc-5snpr [743.801082ms] - Dec 14 17:16:10.456: INFO: Created: latency-svc-2xzfw - Dec 14 17:16:10.489: INFO: Got endpoints: latency-svc-jrjmt [748.813722ms] - Dec 14 17:16:10.515: INFO: Created: latency-svc-fh6tz - Dec 14 17:16:10.546: INFO: Got endpoints: latency-svc-w5l7k [743.821643ms] - Dec 14 17:16:10.562: INFO: Created: latency-svc-q6w8r - Dec 14 17:16:10.590: INFO: Got endpoints: latency-svc-jhhpb [744.90311ms] - Dec 14 17:16:10.611: INFO: Created: latency-svc-wkh79 - Dec 14 17:16:10.640: INFO: Got endpoints: latency-svc-8pnkx [745.236967ms] - Dec 14 17:16:10.672: INFO: Created: latency-svc-kq8lz - Dec 14 17:16:10.690: INFO: Got endpoints: latency-svc-hhvql [747.853666ms] - Dec 14 17:16:10.702: INFO: Created: latency-svc-dz6jx - Dec 14 17:16:10.741: INFO: Got endpoints: latency-svc-29jd4 [736.818536ms] - Dec 14 17:16:10.756: INFO: Created: latency-svc-pqphm - Dec 14 17:16:10.797: INFO: Got endpoints: latency-svc-ntd4n [753.356684ms] - Dec 14 17:16:10.841: INFO: Got endpoints: latency-svc-z8zx6 [750.221257ms] - Dec 14 17:16:10.896: INFO: Got endpoints: latency-svc-vbb4n [751.164319ms] - Dec 14 17:16:10.941: INFO: Got endpoints: latency-svc-nhcqj [741.734122ms] - Dec 14 17:16:10.994: INFO: Got endpoints: latency-svc-sl6j8 [754.540575ms] - Dec 14 17:16:11.042: INFO: Got endpoints: latency-svc-sn262 [751.264467ms] - Dec 14 17:16:11.093: INFO: Got endpoints: latency-svc-gglvr [752.072589ms] - Dec 14 17:16:11.149: INFO: Got endpoints: latency-svc-pp26q [757.885809ms] - Dec 14 17:16:11.195: INFO: Got endpoints: latency-svc-2xzfw [757.382418ms] - Dec 14 17:16:11.239: INFO: Got endpoints: latency-svc-fh6tz [750.00636ms] - Dec 14 17:16:11.292: INFO: Got endpoints: latency-svc-q6w8r [745.125347ms] - Dec 14 17:16:11.340: INFO: Got endpoints: latency-svc-wkh79 [750.516841ms] - Dec 14 17:16:11.391: INFO: Got endpoints: latency-svc-kq8lz [750.885554ms] - Dec 14 17:16:11.441: INFO: Got endpoints: latency-svc-dz6jx [751.218228ms] - Dec 14 17:16:11.494: INFO: Got endpoints: latency-svc-pqphm [752.809407ms] - Dec 14 17:16:11.494: INFO: Latencies: [31.396245ms 141.208819ms 141.994896ms 163.230592ms 165.788645ms 182.137895ms 183.797748ms 196.846151ms 203.596618ms 221.701232ms 234.213126ms 240.271249ms 246.711469ms 253.527213ms 255.97453ms 257.992284ms 261.528179ms 262.079176ms 273.54668ms 281.270241ms 281.859048ms 283.741899ms 284.665018ms 287.891065ms 297.766123ms 299.208111ms 304.480559ms 309.272446ms 309.409449ms 310.229673ms 317.421903ms 318.867562ms 319.346279ms 329.47754ms 332.274215ms 336.006585ms 336.170935ms 336.421299ms 337.007557ms 341.811913ms 342.52653ms 345.118266ms 350.53158ms 353.648852ms 354.620515ms 355.575227ms 362.204865ms 365.509196ms 376.852681ms 382.685597ms 394.641869ms 417.611771ms 464.683146ms 480.530728ms 508.218847ms 525.76334ms 574.349617ms 580.957445ms 617.396396ms 653.692434ms 661.04851ms 668.502784ms 678.87874ms 681.836381ms 685.342031ms 700.445312ms 703.945132ms 706.774845ms 707.476595ms 707.882394ms 713.327205ms 715.473646ms 715.563427ms 717.573346ms 720.144517ms 723.16495ms 724.981018ms 725.704781ms 726.069733ms 726.832839ms 732.385935ms 733.226417ms 734.362717ms 734.575075ms 736.818536ms 739.198718ms 739.691885ms 740.24763ms 740.911638ms 741.734122ms 742.61698ms 743.022723ms 743.416321ms 743.801082ms 743.821643ms 743.86906ms 743.928118ms 743.985635ms 744.16682ms 744.226084ms 744.555911ms 744.588083ms 744.640515ms 744.678451ms 744.700531ms 744.90311ms 744.925734ms 745.125347ms 745.206213ms 745.236967ms 745.474333ms 745.64416ms 745.69114ms 746.199526ms 746.348404ms 747.051974ms 747.40927ms 747.761301ms 747.853666ms 748.11265ms 748.353189ms 748.388158ms 748.401585ms 748.536241ms 748.5896ms 748.744933ms 748.78858ms 748.813722ms 749.373759ms 749.506755ms 749.899751ms 750.00636ms 750.067526ms 750.121135ms 750.205498ms 750.221257ms 750.253503ms 750.461963ms 750.516841ms 750.692442ms 750.740785ms 750.885554ms 750.960083ms 751.164319ms 751.198596ms 751.210027ms 751.218228ms 751.264467ms 751.566716ms 751.667485ms 752.072589ms 752.1058ms 752.146478ms 752.809407ms 752.865812ms 753.13947ms 753.171167ms 753.313339ms 753.356684ms 753.495156ms 753.535874ms 753.905958ms 754.138026ms 754.267231ms 754.540575ms 754.66886ms 754.929511ms 755.05949ms 755.622399ms 756.067844ms 756.148074ms 756.5254ms 756.855239ms 757.148494ms 757.382418ms 757.885809ms 757.992513ms 758.251774ms 758.569214ms 758.875675ms 759.234702ms 759.466289ms 759.627201ms 762.16168ms 764.25947ms 765.06147ms 765.192545ms 768.221977ms 771.193668ms 771.7579ms 773.339119ms 779.433421ms 781.110117ms 782.124819ms 786.227668ms 786.326179ms 791.337642ms 793.324979ms 823.091477ms 931.482307ms] - Dec 14 17:16:11.494: INFO: 50 %ile: 744.555911ms - Dec 14 17:16:11.494: INFO: 90 %ile: 759.234702ms - Dec 14 17:16:11.494: INFO: 99 %ile: 823.091477ms - Dec 14 17:16:11.494: INFO: Total sample count: 200 - [AfterEach] [sig-network] Service endpoints latency + STEP: Creating a kubernetes client 02/25/23 11:43:30.014 + Feb 25 11:43:30.014: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename namespaces 02/25/23 11:43:30.019 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:43:30.073 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:43:30.08 + [It] should ensure that all services are removed when a namespace is deleted [Conformance] + test/e2e/apimachinery/namespace.go:250 + STEP: Creating a test namespace 02/25/23 11:43:30.084 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:43:30.19 + STEP: Creating a service in the namespace 02/25/23 11:43:30.201 + STEP: Deleting the namespace 02/25/23 11:43:30.225 + STEP: Waiting for the namespace to be removed. 02/25/23 11:43:30.241 + STEP: Recreating the namespace 02/25/23 11:43:36.25 + STEP: Verifying there is no service in the namespace 02/25/23 11:43:36.282 + [AfterEach] [sig-api-machinery] Namespaces [Serial] test/e2e/framework/framework.go:187 - Dec 14 17:16:11.494: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "svc-latency-4007" for this suite. 12/14/22 17:16:11.507 + Feb 25 11:43:36.287: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "namespaces-1616" for this suite. 02/25/23 11:43:36.304 + STEP: Destroying namespace "nsdeletetest-3035" for this suite. 02/25/23 11:43:36.316 + Feb 25 11:43:36.322: INFO: Namespace nsdeletetest-3035 was already deleted + STEP: Destroying namespace "nsdeletetest-9683" for this suite. 02/25/23 11:43:36.322 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSSS +SSSSSSSSSSSSSSSSS ------------------------------ [sig-node] Variable Expansion - should allow substituting values in a container's args [NodeConformance] [Conformance] - test/e2e/common/node/expansion.go:91 + should allow substituting values in a volume subpath [Conformance] + test/e2e/common/node/expansion.go:111 [BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:16:11.52 -Dec 14 17:16:11.520: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename var-expansion 12/14/22 17:16:11.522 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:16:11.547 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:16:11.555 -[It] should allow substituting values in a container's args [NodeConformance] [Conformance] - test/e2e/common/node/expansion.go:91 -STEP: Creating a pod to test substitution in container's args 12/14/22 17:16:11.558 -Dec 14 17:16:11.570: INFO: Waiting up to 5m0s for pod "var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87" in namespace "var-expansion-5645" to be "Succeeded or Failed" -Dec 14 17:16:11.574: INFO: Pod "var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87": Phase="Pending", Reason="", readiness=false. Elapsed: 4.355677ms -Dec 14 17:16:13.583: INFO: Pod "var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013518601s -Dec 14 17:16:15.582: INFO: Pod "var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012178438s -Dec 14 17:16:17.596: INFO: Pod "var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87": Phase="Pending", Reason="", readiness=false. Elapsed: 6.026741268s -Dec 14 17:16:19.581: INFO: Pod "var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.011469904s -STEP: Saw pod success 12/14/22 17:16:19.581 -Dec 14 17:16:19.582: INFO: Pod "var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87" satisfied condition "Succeeded or Failed" -Dec 14 17:16:19.587: INFO: Trying to get logs from node kiedai4iakae-3 pod var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87 container dapi-container: -STEP: delete the pod 12/14/22 17:16:19.622 -Dec 14 17:16:19.648: INFO: Waiting for pod var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87 to disappear -Dec 14 17:16:19.654: INFO: Pod var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87 no longer exists +STEP: Creating a kubernetes client 02/25/23 11:43:36.339 +Feb 25 11:43:36.339: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename var-expansion 02/25/23 11:43:36.341 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:43:36.375 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:43:36.382 +[It] should allow substituting values in a volume subpath [Conformance] + test/e2e/common/node/expansion.go:111 +STEP: Creating a pod to test substitution in volume subpath 02/25/23 11:43:36.39 +Feb 25 11:43:36.403: INFO: Waiting up to 5m0s for pod "var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75" in namespace "var-expansion-428" to be "Succeeded or Failed" +Feb 25 11:43:36.409: INFO: Pod "var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75": Phase="Pending", Reason="", readiness=false. Elapsed: 5.584842ms +Feb 25 11:43:38.417: INFO: Pod "var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013464861s +Feb 25 11:43:40.416: INFO: Pod "var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75": Phase="Pending", Reason="", readiness=false. Elapsed: 4.0125311s +Feb 25 11:43:42.415: INFO: Pod "var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.011411109s +STEP: Saw pod success 02/25/23 11:43:42.415 +Feb 25 11:43:42.415: INFO: Pod "var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75" satisfied condition "Succeeded or Failed" +Feb 25 11:43:42.419: INFO: Trying to get logs from node diek3dua9gay-3 pod var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75 container dapi-container: +STEP: delete the pod 02/25/23 11:43:42.429 +Feb 25 11:43:42.448: INFO: Waiting for pod var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75 to disappear +Feb 25 11:43:42.452: INFO: Pod var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75 no longer exists [AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 -Dec 14 17:16:19.654: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "var-expansion-5645" for this suite. 12/14/22 17:16:19.661 -{"msg":"PASSED [sig-node] Variable Expansion should allow substituting values in a container's args [NodeConformance] [Conformance]","completed":361,"skipped":6691,"failed":0} +Feb 25 11:43:42.452: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "var-expansion-428" for this suite. 02/25/23 11:43:42.46 +{"msg":"PASSED [sig-node] Variable Expansion should allow substituting values in a volume subpath [Conformance]","completed":361,"skipped":6694,"failed":0} ------------------------------ -• [SLOW TEST] [8.193 seconds] +• [SLOW TEST] [6.131 seconds] [sig-node] Variable Expansion test/e2e/common/node/framework.go:23 - should allow substituting values in a container's args [NodeConformance] [Conformance] - test/e2e/common/node/expansion.go:91 + should allow substituting values in a volume subpath [Conformance] + test/e2e/common/node/expansion.go:111 Begin Captured GinkgoWriter Output >> [BeforeEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:16:11.52 - Dec 14 17:16:11.520: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename var-expansion 12/14/22 17:16:11.522 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:16:11.547 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:16:11.555 - [It] should allow substituting values in a container's args [NodeConformance] [Conformance] - test/e2e/common/node/expansion.go:91 - STEP: Creating a pod to test substitution in container's args 12/14/22 17:16:11.558 - Dec 14 17:16:11.570: INFO: Waiting up to 5m0s for pod "var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87" in namespace "var-expansion-5645" to be "Succeeded or Failed" - Dec 14 17:16:11.574: INFO: Pod "var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87": Phase="Pending", Reason="", readiness=false. Elapsed: 4.355677ms - Dec 14 17:16:13.583: INFO: Pod "var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013518601s - Dec 14 17:16:15.582: INFO: Pod "var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87": Phase="Pending", Reason="", readiness=false. Elapsed: 4.012178438s - Dec 14 17:16:17.596: INFO: Pod "var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87": Phase="Pending", Reason="", readiness=false. Elapsed: 6.026741268s - Dec 14 17:16:19.581: INFO: Pod "var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87": Phase="Succeeded", Reason="", readiness=false. Elapsed: 8.011469904s - STEP: Saw pod success 12/14/22 17:16:19.581 - Dec 14 17:16:19.582: INFO: Pod "var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87" satisfied condition "Succeeded or Failed" - Dec 14 17:16:19.587: INFO: Trying to get logs from node kiedai4iakae-3 pod var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87 container dapi-container: - STEP: delete the pod 12/14/22 17:16:19.622 - Dec 14 17:16:19.648: INFO: Waiting for pod var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87 to disappear - Dec 14 17:16:19.654: INFO: Pod var-expansion-8fed09fb-212c-4033-8dca-d421b1ad5c87 no longer exists + STEP: Creating a kubernetes client 02/25/23 11:43:36.339 + Feb 25 11:43:36.339: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename var-expansion 02/25/23 11:43:36.341 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:43:36.375 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:43:36.382 + [It] should allow substituting values in a volume subpath [Conformance] + test/e2e/common/node/expansion.go:111 + STEP: Creating a pod to test substitution in volume subpath 02/25/23 11:43:36.39 + Feb 25 11:43:36.403: INFO: Waiting up to 5m0s for pod "var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75" in namespace "var-expansion-428" to be "Succeeded or Failed" + Feb 25 11:43:36.409: INFO: Pod "var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75": Phase="Pending", Reason="", readiness=false. Elapsed: 5.584842ms + Feb 25 11:43:38.417: INFO: Pod "var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75": Phase="Pending", Reason="", readiness=false. Elapsed: 2.013464861s + Feb 25 11:43:40.416: INFO: Pod "var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75": Phase="Pending", Reason="", readiness=false. Elapsed: 4.0125311s + Feb 25 11:43:42.415: INFO: Pod "var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75": Phase="Succeeded", Reason="", readiness=false. Elapsed: 6.011411109s + STEP: Saw pod success 02/25/23 11:43:42.415 + Feb 25 11:43:42.415: INFO: Pod "var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75" satisfied condition "Succeeded or Failed" + Feb 25 11:43:42.419: INFO: Trying to get logs from node diek3dua9gay-3 pod var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75 container dapi-container: + STEP: delete the pod 02/25/23 11:43:42.429 + Feb 25 11:43:42.448: INFO: Waiting for pod var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75 to disappear + Feb 25 11:43:42.452: INFO: Pod var-expansion-1e1df440-eeac-4fc7-91c5-766f05f7ee75 no longer exists [AfterEach] [sig-node] Variable Expansion test/e2e/framework/framework.go:187 - Dec 14 17:16:19.654: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "var-expansion-5645" for this suite. 12/14/22 17:16:19.661 + Feb 25 11:43:42.452: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "var-expansion-428" for this suite. 02/25/23 11:43:42.46 << End Captured GinkgoWriter Output ------------------------------ -SSSSSSSSSSS +S ------------------------------ -[sig-network] DNS - should provide DNS for ExternalName services [Conformance] - test/e2e/network/dns.go:333 -[BeforeEach] [sig-network] DNS +[sig-node] NoExecuteTaintManager Multiple Pods [Serial] + evicts pods with minTolerationSeconds [Disruptive] [Conformance] + test/e2e/node/taints.go:420 +[BeforeEach] [sig-node] NoExecuteTaintManager Multiple Pods [Serial] test/e2e/framework/framework.go:186 -STEP: Creating a kubernetes client 12/14/22 17:16:19.718 -Dec 14 17:16:19.719: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 -STEP: Building a namespace api object, basename dns 12/14/22 17:16:19.721 -STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:16:19.757 -STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:16:19.761 -[It] should provide DNS for ExternalName services [Conformance] - test/e2e/network/dns.go:333 -STEP: Creating a test externalName service 12/14/22 17:16:19.764 -STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-3425.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local; sleep 1; done - 12/14/22 17:16:19.777 -STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-3425.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local; sleep 1; done - 12/14/22 17:16:19.778 -STEP: creating a pod to probe DNS 12/14/22 17:16:19.778 -STEP: submitting the pod to kubernetes 12/14/22 17:16:19.778 -Dec 14 17:16:19.858: INFO: Waiting up to 15m0s for pod "dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35" in namespace "dns-3425" to be "running" -Dec 14 17:16:19.871: INFO: Pod "dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35": Phase="Pending", Reason="", readiness=false. Elapsed: 12.501263ms -Dec 14 17:16:21.881: INFO: Pod "dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022358064s -Dec 14 17:16:23.887: INFO: Pod "dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35": Phase="Pending", Reason="", readiness=false. Elapsed: 4.028532385s -Dec 14 17:16:25.878: INFO: Pod "dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35": Phase="Pending", Reason="", readiness=false. Elapsed: 6.019441727s -Dec 14 17:16:27.878: INFO: Pod "dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35": Phase="Running", Reason="", readiness=true. Elapsed: 8.019606678s -Dec 14 17:16:27.878: INFO: Pod "dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35" satisfied condition "running" -STEP: retrieving the pod 12/14/22 17:16:27.878 -STEP: looking for the results for each expected name from probers 12/14/22 17:16:27.883 -Dec 14 17:16:27.893: INFO: DNS probes using dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35 succeeded - -STEP: deleting the pod 12/14/22 17:16:27.893 -STEP: changing the externalName to bar.example.com 12/14/22 17:16:27.963 -STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-3425.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local; sleep 1; done - 12/14/22 17:16:27.979 -STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-3425.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local; sleep 1; done - 12/14/22 17:16:27.979 -STEP: creating a second pod to probe DNS 12/14/22 17:16:27.979 -STEP: submitting the pod to kubernetes 12/14/22 17:16:27.979 -Dec 14 17:16:27.986: INFO: Waiting up to 15m0s for pod "dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40" in namespace "dns-3425" to be "running" -Dec 14 17:16:27.990: INFO: Pod "dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40": Phase="Pending", Reason="", readiness=false. Elapsed: 3.541461ms -Dec 14 17:16:29.998: INFO: Pod "dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012169826s -Dec 14 17:16:31.995: INFO: Pod "dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40": Phase="Pending", Reason="", readiness=false. Elapsed: 4.009201109s -Dec 14 17:16:33.996: INFO: Pod "dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40": Phase="Running", Reason="", readiness=true. Elapsed: 6.009988082s -Dec 14 17:16:33.996: INFO: Pod "dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40" satisfied condition "running" -STEP: retrieving the pod 12/14/22 17:16:33.996 -STEP: looking for the results for each expected name from probers 12/14/22 17:16:34 -Dec 14 17:16:34.008: INFO: File wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. -' instead of 'bar.example.com.' -Dec 14 17:16:34.017: INFO: File jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. -' instead of 'bar.example.com.' -Dec 14 17:16:34.017: INFO: Lookups using dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 failed for: [wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local] - -Dec 14 17:16:39.025: INFO: File wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. -' instead of 'bar.example.com.' -Dec 14 17:16:39.030: INFO: File jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. -' instead of 'bar.example.com.' -Dec 14 17:16:39.030: INFO: Lookups using dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 failed for: [wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local] - -Dec 14 17:16:44.023: INFO: File wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. -' instead of 'bar.example.com.' -Dec 14 17:16:44.026: INFO: File jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. -' instead of 'bar.example.com.' -Dec 14 17:16:44.026: INFO: Lookups using dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 failed for: [wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local] - -Dec 14 17:16:49.023: INFO: File wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. -' instead of 'bar.example.com.' -Dec 14 17:16:49.029: INFO: File jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. -' instead of 'bar.example.com.' -Dec 14 17:16:49.029: INFO: Lookups using dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 failed for: [wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local] - -Dec 14 17:16:54.024: INFO: File wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. -' instead of 'bar.example.com.' -Dec 14 17:16:54.027: INFO: File jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. -' instead of 'bar.example.com.' -Dec 14 17:16:54.027: INFO: Lookups using dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 failed for: [wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local] - -Dec 14 17:16:59.031: INFO: DNS probes using dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 succeeded - -STEP: deleting the pod 12/14/22 17:16:59.031 -STEP: changing the service to type=ClusterIP 12/14/22 17:16:59.05 -STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-3425.svc.cluster.local A > /results/wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local; sleep 1; done - 12/14/22 17:16:59.121 -STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-3425.svc.cluster.local A > /results/jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local; sleep 1; done - 12/14/22 17:16:59.121 -STEP: creating a third pod to probe DNS 12/14/22 17:16:59.121 -STEP: submitting the pod to kubernetes 12/14/22 17:16:59.127 -Dec 14 17:16:59.146: INFO: Waiting up to 15m0s for pod "dns-test-91ff6788-fbe8-41ed-add5-e7041746efab" in namespace "dns-3425" to be "running" -Dec 14 17:16:59.191: INFO: Pod "dns-test-91ff6788-fbe8-41ed-add5-e7041746efab": Phase="Pending", Reason="", readiness=false. Elapsed: 44.896699ms -Dec 14 17:17:01.199: INFO: Pod "dns-test-91ff6788-fbe8-41ed-add5-e7041746efab": Phase="Pending", Reason="", readiness=false. Elapsed: 2.052294609s -Dec 14 17:17:03.197: INFO: Pod "dns-test-91ff6788-fbe8-41ed-add5-e7041746efab": Phase="Running", Reason="", readiness=true. Elapsed: 4.050819353s -Dec 14 17:17:03.197: INFO: Pod "dns-test-91ff6788-fbe8-41ed-add5-e7041746efab" satisfied condition "running" -STEP: retrieving the pod 12/14/22 17:17:03.198 -STEP: looking for the results for each expected name from probers 12/14/22 17:17:03.202 -Dec 14 17:17:03.213: INFO: DNS probes using dns-test-91ff6788-fbe8-41ed-add5-e7041746efab succeeded - -STEP: deleting the pod 12/14/22 17:17:03.213 -STEP: deleting the test externalName service 12/14/22 17:17:03.23 -[AfterEach] [sig-network] DNS +STEP: Creating a kubernetes client 02/25/23 11:43:42.47 +Feb 25 11:43:42.470: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 +STEP: Building a namespace api object, basename taint-multiple-pods 02/25/23 11:43:42.472 +STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:43:42.498 +STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:43:42.503 +[BeforeEach] [sig-node] NoExecuteTaintManager Multiple Pods [Serial] + test/e2e/node/taints.go:348 +Feb 25 11:43:42.506: INFO: Waiting up to 1m0s for all nodes to be ready +Feb 25 11:44:42.541: INFO: Waiting for terminating namespaces to be deleted... +[It] evicts pods with minTolerationSeconds [Disruptive] [Conformance] + test/e2e/node/taints.go:420 +Feb 25 11:44:42.547: INFO: Starting informer... +STEP: Starting pods... 02/25/23 11:44:42.548 +Feb 25 11:44:42.777: INFO: Pod1 is running on diek3dua9gay-3. Tainting Node +Feb 25 11:44:42.989: INFO: Waiting up to 5m0s for pod "taint-eviction-b1" in namespace "taint-multiple-pods-1906" to be "running" +Feb 25 11:44:43.000: INFO: Pod "taint-eviction-b1": Phase="Pending", Reason="", readiness=false. Elapsed: 10.536489ms +Feb 25 11:44:45.015: INFO: Pod "taint-eviction-b1": Phase="Running", Reason="", readiness=true. Elapsed: 2.025781436s +Feb 25 11:44:45.015: INFO: Pod "taint-eviction-b1" satisfied condition "running" +Feb 25 11:44:45.015: INFO: Waiting up to 5m0s for pod "taint-eviction-b2" in namespace "taint-multiple-pods-1906" to be "running" +Feb 25 11:44:45.025: INFO: Pod "taint-eviction-b2": Phase="Pending", Reason="", readiness=false. Elapsed: 9.466131ms +Feb 25 11:44:47.031: INFO: Pod "taint-eviction-b2": Phase="Running", Reason="", readiness=true. Elapsed: 2.01560889s +Feb 25 11:44:47.031: INFO: Pod "taint-eviction-b2" satisfied condition "running" +Feb 25 11:44:47.031: INFO: Pod2 is running on diek3dua9gay-3. Tainting Node +STEP: Trying to apply a taint on the Node 02/25/23 11:44:47.031 +STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 02/25/23 11:44:47.051 +STEP: Waiting for Pod1 and Pod2 to be deleted 02/25/23 11:44:47.062 +Feb 25 11:44:53.308: INFO: Noticed Pod "taint-eviction-b1" gets evicted. +Feb 25 11:45:13.355: INFO: Noticed Pod "taint-eviction-b2" gets evicted. +STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 02/25/23 11:45:13.382 +[AfterEach] [sig-node] NoExecuteTaintManager Multiple Pods [Serial] test/e2e/framework/framework.go:187 -Dec 14 17:17:03.261: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready -STEP: Destroying namespace "dns-3425" for this suite. 12/14/22 17:17:03.268 -{"msg":"PASSED [sig-network] DNS should provide DNS for ExternalName services [Conformance]","completed":362,"skipped":6702,"failed":0} +Feb 25 11:45:13.390: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready +STEP: Destroying namespace "taint-multiple-pods-1906" for this suite. 02/25/23 11:45:13.402 +{"msg":"PASSED [sig-node] NoExecuteTaintManager Multiple Pods [Serial] evicts pods with minTolerationSeconds [Disruptive] [Conformance]","completed":362,"skipped":6695,"failed":0} ------------------------------ -• [SLOW TEST] [43.559 seconds] -[sig-network] DNS -test/e2e/network/common/framework.go:23 - should provide DNS for ExternalName services [Conformance] - test/e2e/network/dns.go:333 +• [SLOW TEST] [90.957 seconds] +[sig-node] NoExecuteTaintManager Multiple Pods [Serial] +test/e2e/node/framework.go:23 + evicts pods with minTolerationSeconds [Disruptive] [Conformance] + test/e2e/node/taints.go:420 Begin Captured GinkgoWriter Output >> - [BeforeEach] [sig-network] DNS + [BeforeEach] [sig-node] NoExecuteTaintManager Multiple Pods [Serial] test/e2e/framework/framework.go:186 - STEP: Creating a kubernetes client 12/14/22 17:16:19.718 - Dec 14 17:16:19.719: INFO: >>> kubeConfig: /tmp/kubeconfig-3507870282 - STEP: Building a namespace api object, basename dns 12/14/22 17:16:19.721 - STEP: Waiting for a default service account to be provisioned in namespace 12/14/22 17:16:19.757 - STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 12/14/22 17:16:19.761 - [It] should provide DNS for ExternalName services [Conformance] - test/e2e/network/dns.go:333 - STEP: Creating a test externalName service 12/14/22 17:16:19.764 - STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-3425.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local; sleep 1; done - 12/14/22 17:16:19.777 - STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-3425.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local; sleep 1; done - 12/14/22 17:16:19.778 - STEP: creating a pod to probe DNS 12/14/22 17:16:19.778 - STEP: submitting the pod to kubernetes 12/14/22 17:16:19.778 - Dec 14 17:16:19.858: INFO: Waiting up to 15m0s for pod "dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35" in namespace "dns-3425" to be "running" - Dec 14 17:16:19.871: INFO: Pod "dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35": Phase="Pending", Reason="", readiness=false. Elapsed: 12.501263ms - Dec 14 17:16:21.881: INFO: Pod "dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35": Phase="Pending", Reason="", readiness=false. Elapsed: 2.022358064s - Dec 14 17:16:23.887: INFO: Pod "dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35": Phase="Pending", Reason="", readiness=false. Elapsed: 4.028532385s - Dec 14 17:16:25.878: INFO: Pod "dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35": Phase="Pending", Reason="", readiness=false. Elapsed: 6.019441727s - Dec 14 17:16:27.878: INFO: Pod "dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35": Phase="Running", Reason="", readiness=true. Elapsed: 8.019606678s - Dec 14 17:16:27.878: INFO: Pod "dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35" satisfied condition "running" - STEP: retrieving the pod 12/14/22 17:16:27.878 - STEP: looking for the results for each expected name from probers 12/14/22 17:16:27.883 - Dec 14 17:16:27.893: INFO: DNS probes using dns-test-eefdb4c4-c516-493e-ab8d-1bc0efcaec35 succeeded - - STEP: deleting the pod 12/14/22 17:16:27.893 - STEP: changing the externalName to bar.example.com 12/14/22 17:16:27.963 - STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-3425.svc.cluster.local CNAME > /results/wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local; sleep 1; done - 12/14/22 17:16:27.979 - STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-3425.svc.cluster.local CNAME > /results/jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local; sleep 1; done - 12/14/22 17:16:27.979 - STEP: creating a second pod to probe DNS 12/14/22 17:16:27.979 - STEP: submitting the pod to kubernetes 12/14/22 17:16:27.979 - Dec 14 17:16:27.986: INFO: Waiting up to 15m0s for pod "dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40" in namespace "dns-3425" to be "running" - Dec 14 17:16:27.990: INFO: Pod "dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40": Phase="Pending", Reason="", readiness=false. Elapsed: 3.541461ms - Dec 14 17:16:29.998: INFO: Pod "dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40": Phase="Pending", Reason="", readiness=false. Elapsed: 2.012169826s - Dec 14 17:16:31.995: INFO: Pod "dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40": Phase="Pending", Reason="", readiness=false. Elapsed: 4.009201109s - Dec 14 17:16:33.996: INFO: Pod "dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40": Phase="Running", Reason="", readiness=true. Elapsed: 6.009988082s - Dec 14 17:16:33.996: INFO: Pod "dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40" satisfied condition "running" - STEP: retrieving the pod 12/14/22 17:16:33.996 - STEP: looking for the results for each expected name from probers 12/14/22 17:16:34 - Dec 14 17:16:34.008: INFO: File wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. - ' instead of 'bar.example.com.' - Dec 14 17:16:34.017: INFO: File jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. - ' instead of 'bar.example.com.' - Dec 14 17:16:34.017: INFO: Lookups using dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 failed for: [wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local] - - Dec 14 17:16:39.025: INFO: File wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. - ' instead of 'bar.example.com.' - Dec 14 17:16:39.030: INFO: File jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. - ' instead of 'bar.example.com.' - Dec 14 17:16:39.030: INFO: Lookups using dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 failed for: [wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local] - - Dec 14 17:16:44.023: INFO: File wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. - ' instead of 'bar.example.com.' - Dec 14 17:16:44.026: INFO: File jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. - ' instead of 'bar.example.com.' - Dec 14 17:16:44.026: INFO: Lookups using dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 failed for: [wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local] - - Dec 14 17:16:49.023: INFO: File wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. - ' instead of 'bar.example.com.' - Dec 14 17:16:49.029: INFO: File jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. - ' instead of 'bar.example.com.' - Dec 14 17:16:49.029: INFO: Lookups using dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 failed for: [wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local] - - Dec 14 17:16:54.024: INFO: File wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. - ' instead of 'bar.example.com.' - Dec 14 17:16:54.027: INFO: File jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local from pod dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 contains 'foo.example.com. - ' instead of 'bar.example.com.' - Dec 14 17:16:54.027: INFO: Lookups using dns-3425/dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 failed for: [wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local] - - Dec 14 17:16:59.031: INFO: DNS probes using dns-test-79f0ff6d-5bca-4218-910b-21b37e499f40 succeeded - - STEP: deleting the pod 12/14/22 17:16:59.031 - STEP: changing the service to type=ClusterIP 12/14/22 17:16:59.05 - STEP: Running these commands on wheezy: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-3425.svc.cluster.local A > /results/wheezy_udp@dns-test-service-3.dns-3425.svc.cluster.local; sleep 1; done - 12/14/22 17:16:59.121 - STEP: Running these commands on jessie: for i in `seq 1 30`; do dig +short dns-test-service-3.dns-3425.svc.cluster.local A > /results/jessie_udp@dns-test-service-3.dns-3425.svc.cluster.local; sleep 1; done - 12/14/22 17:16:59.121 - STEP: creating a third pod to probe DNS 12/14/22 17:16:59.121 - STEP: submitting the pod to kubernetes 12/14/22 17:16:59.127 - Dec 14 17:16:59.146: INFO: Waiting up to 15m0s for pod "dns-test-91ff6788-fbe8-41ed-add5-e7041746efab" in namespace "dns-3425" to be "running" - Dec 14 17:16:59.191: INFO: Pod "dns-test-91ff6788-fbe8-41ed-add5-e7041746efab": Phase="Pending", Reason="", readiness=false. Elapsed: 44.896699ms - Dec 14 17:17:01.199: INFO: Pod "dns-test-91ff6788-fbe8-41ed-add5-e7041746efab": Phase="Pending", Reason="", readiness=false. Elapsed: 2.052294609s - Dec 14 17:17:03.197: INFO: Pod "dns-test-91ff6788-fbe8-41ed-add5-e7041746efab": Phase="Running", Reason="", readiness=true. Elapsed: 4.050819353s - Dec 14 17:17:03.197: INFO: Pod "dns-test-91ff6788-fbe8-41ed-add5-e7041746efab" satisfied condition "running" - STEP: retrieving the pod 12/14/22 17:17:03.198 - STEP: looking for the results for each expected name from probers 12/14/22 17:17:03.202 - Dec 14 17:17:03.213: INFO: DNS probes using dns-test-91ff6788-fbe8-41ed-add5-e7041746efab succeeded - - STEP: deleting the pod 12/14/22 17:17:03.213 - STEP: deleting the test externalName service 12/14/22 17:17:03.23 - [AfterEach] [sig-network] DNS + STEP: Creating a kubernetes client 02/25/23 11:43:42.47 + Feb 25 11:43:42.470: INFO: >>> kubeConfig: /tmp/kubeconfig-2917274343 + STEP: Building a namespace api object, basename taint-multiple-pods 02/25/23 11:43:42.472 + STEP: Waiting for a default service account to be provisioned in namespace 02/25/23 11:43:42.498 + STEP: Waiting for kube-root-ca.crt to be provisioned in namespace 02/25/23 11:43:42.503 + [BeforeEach] [sig-node] NoExecuteTaintManager Multiple Pods [Serial] + test/e2e/node/taints.go:348 + Feb 25 11:43:42.506: INFO: Waiting up to 1m0s for all nodes to be ready + Feb 25 11:44:42.541: INFO: Waiting for terminating namespaces to be deleted... + [It] evicts pods with minTolerationSeconds [Disruptive] [Conformance] + test/e2e/node/taints.go:420 + Feb 25 11:44:42.547: INFO: Starting informer... + STEP: Starting pods... 02/25/23 11:44:42.548 + Feb 25 11:44:42.777: INFO: Pod1 is running on diek3dua9gay-3. Tainting Node + Feb 25 11:44:42.989: INFO: Waiting up to 5m0s for pod "taint-eviction-b1" in namespace "taint-multiple-pods-1906" to be "running" + Feb 25 11:44:43.000: INFO: Pod "taint-eviction-b1": Phase="Pending", Reason="", readiness=false. Elapsed: 10.536489ms + Feb 25 11:44:45.015: INFO: Pod "taint-eviction-b1": Phase="Running", Reason="", readiness=true. Elapsed: 2.025781436s + Feb 25 11:44:45.015: INFO: Pod "taint-eviction-b1" satisfied condition "running" + Feb 25 11:44:45.015: INFO: Waiting up to 5m0s for pod "taint-eviction-b2" in namespace "taint-multiple-pods-1906" to be "running" + Feb 25 11:44:45.025: INFO: Pod "taint-eviction-b2": Phase="Pending", Reason="", readiness=false. Elapsed: 9.466131ms + Feb 25 11:44:47.031: INFO: Pod "taint-eviction-b2": Phase="Running", Reason="", readiness=true. Elapsed: 2.01560889s + Feb 25 11:44:47.031: INFO: Pod "taint-eviction-b2" satisfied condition "running" + Feb 25 11:44:47.031: INFO: Pod2 is running on diek3dua9gay-3. Tainting Node + STEP: Trying to apply a taint on the Node 02/25/23 11:44:47.031 + STEP: verifying the node has the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 02/25/23 11:44:47.051 + STEP: Waiting for Pod1 and Pod2 to be deleted 02/25/23 11:44:47.062 + Feb 25 11:44:53.308: INFO: Noticed Pod "taint-eviction-b1" gets evicted. + Feb 25 11:45:13.355: INFO: Noticed Pod "taint-eviction-b2" gets evicted. + STEP: verifying the node doesn't have the taint kubernetes.io/e2e-evict-taint-key=evictTaintVal:NoExecute 02/25/23 11:45:13.382 + [AfterEach] [sig-node] NoExecuteTaintManager Multiple Pods [Serial] test/e2e/framework/framework.go:187 - Dec 14 17:17:03.261: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready - STEP: Destroying namespace "dns-3425" for this suite. 12/14/22 17:17:03.268 + Feb 25 11:45:13.390: INFO: Waiting up to 3m0s for all (but 0) nodes to be ready + STEP: Destroying namespace "taint-multiple-pods-1906" for this suite. 02/25/23 11:45:13.402 << End Captured GinkgoWriter Output ------------------------------ -SS +SSSSSSSSS ------------------------------ [SynchronizedAfterSuite] test/e2e/e2e.go:87 [SynchronizedAfterSuite] TOP-LEVEL test/e2e/e2e.go:87 {"msg":"Test Suite completed","completed":362,"skipped":6704,"failed":0} -Dec 14 17:17:03.278: INFO: Running AfterSuite actions on all nodes -Dec 14 17:17:03.278: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage.glob..func20.2 -Dec 14 17:17:03.278: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage.glob..func10.2 -Dec 14 17:17:03.279: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage.glob..func9.2 -Dec 14 17:17:03.279: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func17.3 -Dec 14 17:17:03.279: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func9.2 -Dec 14 17:17:03.279: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func4.2 -Dec 14 17:17:03.279: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func1.3 +Feb 25 11:45:13.433: INFO: Running AfterSuite actions on all nodes +Feb 25 11:45:13.433: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage.glob..func20.2 +Feb 25 11:45:13.433: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage.glob..func10.2 +Feb 25 11:45:13.433: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage.glob..func9.2 +Feb 25 11:45:13.433: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func17.3 +Feb 25 11:45:13.433: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func9.2 +Feb 25 11:45:13.433: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func4.2 +Feb 25 11:45:13.433: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func1.3 [SynchronizedAfterSuite] TOP-LEVEL test/e2e/e2e.go:87 -Dec 14 17:17:03.279: INFO: Running AfterSuite actions on node 1 -Dec 14 17:17:03.279: INFO: Skipping dumping logs from cluster +Feb 25 11:45:13.434: INFO: Running AfterSuite actions on node 1 +Feb 25 11:45:13.434: INFO: Skipping dumping logs from cluster ------------------------------ [SynchronizedAfterSuite] PASSED [0.000 seconds] [SynchronizedAfterSuite] @@ -32154,18 +33102,18 @@ test/e2e/e2e.go:87 Begin Captured GinkgoWriter Output >> [SynchronizedAfterSuite] TOP-LEVEL test/e2e/e2e.go:87 - Dec 14 17:17:03.278: INFO: Running AfterSuite actions on all nodes - Dec 14 17:17:03.278: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage.glob..func20.2 - Dec 14 17:17:03.278: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage.glob..func10.2 - Dec 14 17:17:03.279: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage.glob..func9.2 - Dec 14 17:17:03.279: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func17.3 - Dec 14 17:17:03.279: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func9.2 - Dec 14 17:17:03.279: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func4.2 - Dec 14 17:17:03.279: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func1.3 + Feb 25 11:45:13.433: INFO: Running AfterSuite actions on all nodes + Feb 25 11:45:13.433: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage.glob..func20.2 + Feb 25 11:45:13.433: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage.glob..func10.2 + Feb 25 11:45:13.433: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage.glob..func9.2 + Feb 25 11:45:13.433: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func17.3 + Feb 25 11:45:13.433: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func9.2 + Feb 25 11:45:13.433: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func4.2 + Feb 25 11:45:13.433: INFO: Running Cleanup Action: k8s.io/kubernetes/test/e2e/storage/vsphere.glob..func1.3 [SynchronizedAfterSuite] TOP-LEVEL test/e2e/e2e.go:87 - Dec 14 17:17:03.279: INFO: Running AfterSuite actions on node 1 - Dec 14 17:17:03.279: INFO: Skipping dumping logs from cluster + Feb 25 11:45:13.434: INFO: Running AfterSuite actions on node 1 + Feb 25 11:45:13.434: INFO: Skipping dumping logs from cluster << End Captured GinkgoWriter Output ------------------------------ [ReportAfterSuite] Kubernetes e2e suite report @@ -32187,7 +33135,7 @@ test/e2e/framework/test_context.go:559 [ReportAfterSuite] TOP-LEVEL test/e2e/framework/test_context.go:559 ------------------------------ -[ReportAfterSuite] PASSED [0.123 seconds] +[ReportAfterSuite] PASSED [0.127 seconds] [ReportAfterSuite] Kubernetes e2e JUnit report test/e2e/framework/test_context.go:559 @@ -32197,11 +33145,11 @@ test/e2e/framework/test_context.go:559 << End Captured GinkgoWriter Output ------------------------------ -Ran 362 of 7066 Specs in 6044.954 seconds +Ran 362 of 7066 Specs in 7110.888 seconds SUCCESS! -- 362 Passed | 0 Failed | 0 Pending | 6704 Skipped PASS -Ginkgo ran 1 suite in 1h40m45.654968821s +Ginkgo ran 1 suite in 1h58m31.687374901s Test Suite Passed You're using deprecated Ginkgo functionality: ============================================= diff --git a/v1.25/alvistack-vagrant-kubernetes/junit_01.xml b/v1.25/alvistack-vagrant-kubernetes/junit_01.xml index 1bcd5dc00e..6a5d6ef01f 100644 --- a/v1.25/alvistack-vagrant-kubernetes/junit_01.xml +++ b/v1.25/alvistack-vagrant-kubernetes/junit_01.xml @@ -1,12 +1,12 @@ - - + + - + @@ -21,20482 +21,20482 @@ - - + + - + - + - + - - + + - + - + - + - + - + - + - + - + - - + - + - + - - - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - - + - + - + - + + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - - + - + - + - + - + - - + - + - + - + + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + + - + - + - + - + - - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + + - + + - + + - + - + + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + + - + - - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - - + + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - - - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + + - + - + + - + - + + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - - + + - + - + - + - + - + - + - + - - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + + - + - + - + - + - + - + - + - + + - + - - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + + + - + - + + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - - + - + - + - - + - + - + - + - - + - + - + - + - + - + - + + - + + - - + - + - + + - + - + - + - + - + - + + - + - + - + - - + + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + + + - + + - + - + - + - + - + - + - + + - + - - + - + - + - + - + - + - + - + - - + - + - - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + + + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - - + - + - + - + - + - + - + - + - + - + + - - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + + - + - + - + - + + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - - + - + - + - + - + - + - - + - + - + - + - - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + + - - + - + + - + - + - + - + - + - + - + - - + - + - + + - + - + - + + - + - + - + - + - + - + - + - - + - + - - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - - - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - - + - - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + + + - + + - + - + - + - + - + - + - + + - + - + + + - + + - + - - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + + - + - + - + + - + - + - + - + - + - + - + - - + - + - + - + - + - + - - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + + - + - - + - + - - + - + - + - + - + - + - + - + - + - - + - + - + - + - - + - + - - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + + - - + + - + - + - + - + - + - + + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + + - + - + + - + + - + - + - + - + - + - + - + - + - + - - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + + + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + + - + - + - + - + - + - + - + - + - + - - + - + - - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + + - + - + - + - - + - + - + - + - + - - + - + - + - - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - - - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - - + - + - + - + - + + - + - + - + - - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - - + - + - - + - + - + + - + - + - + - + - + - + - + - - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + + - + - + - - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - - + - + + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - - + - - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - - + - - + - + - + - + - + - + - + - + - + - + - - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - - + - + - - + - + - + - - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + + - + - + - + - + - + - + - + - + - + - + - - - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + + - + - + - + - - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + + - + - + - + - + + - + - + - - + - + - + - - + - + - + - + - + - + - - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + + - + - + - + - - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - - - + - + + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - - + + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + + - + - + - - + - + - + - + - + - + - - + - + + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + + - + + - + - + - - + + + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + + + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + + - + + - + - + - + - + - + - + - + - + - + + - + + - + - + - + - + - + - + - + - + - + - + + - + - + - + - - + - + - + - - + - + - + - - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + + - + - + + - + - + - - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - - + - + - - + - + + - + + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - - + - + - + - - + - + - + - + - + - + - + + - + - + - + - + - + - - + - - + - + - + - - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + + - + - + + - + - + - - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + + + - + - + - + - - + - + - + - + - + - + - + - + - + - + + - - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - - + - + - + - + - + - - - + - - + - - - + - + - - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - - + + - + + - + - + - + - + - + - + - + - + + - + - + - + + - + - + - + + - + - + - - + - + + - + - - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - - - + - - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + + - + - + - + - + - + - + - - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - - + - - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + + - + - + - + - + - + - + + + - + - - + - + + - + - + + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - - + - + - + - + - + + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - - + - + - + - + + - + + - + - + - + - + - - + + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - - + - + - + + - + - + - + - + - + + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + + + - + - + - + - + - + - + - - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + + - + - + - + - + - - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - - - + - - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - - + - + - + + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + + - + - + - - + - + - + - + - + - + - + - + - - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - - + - - + - + - + - + - - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - - + - + - + - + - - + - + - + - + - + - + - + - + + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + + - - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + + - + - - + - + - + + - + - + - + - + - + - + - + - + - + - + + - + - + - - + - + - - - + - + - + - + - + - + - + + - + - + - + - + + - + + - + - + - + - + - + - + - + - + - + - + - + + - + - + + - + - + - + + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + + - + - + + - + - + - + - + - + - + - - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - - + - + - + - + - - + - + - + - - + - + + - + - + - + + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + + + - + - + - + - + - - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - - + - + - + + - + - + - + - + - + - + - + - + + - + - + - - + - + + - + - - + - + - + - + - + - + - + - + + - - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + - - + - + - + + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - - + - + - + - + - + - + - + - + - + - + + - + - + - - + - + - + - + - + - + - + - + - - - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - - + - + - + - + - - + - + - + - + - + + - - + - + - + - - - + - + - + - + + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + - + + - + - + - + + - + - + - + - + - + - + - + - + - + - + - + - + - + - - + - + - + - + + - + + - + - + - + - + - + - + - - + - + - - + + \ No newline at end of file